var/home/core/zuul-output/0000755000175000017500000000000015072351570014532 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015072356556015507 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log0000644000000000000000004550761515072356550017722 0ustar rootrootOct 11 03:51:10 crc systemd[1]: Starting Kubernetes Kubelet... Oct 11 03:51:10 crc restorecon[4675]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 11 03:51:10 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 11 03:51:11 crc restorecon[4675]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Oct 11 03:51:11 crc restorecon[4675]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Oct 11 03:51:12 crc kubenswrapper[4967]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 11 03:51:12 crc kubenswrapper[4967]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Oct 11 03:51:12 crc kubenswrapper[4967]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 11 03:51:12 crc kubenswrapper[4967]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 11 03:51:12 crc kubenswrapper[4967]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Oct 11 03:51:12 crc kubenswrapper[4967]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.562011 4967 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567565 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567599 4967 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567611 4967 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567621 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567633 4967 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567643 4967 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567653 4967 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567662 4967 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567672 4967 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567682 4967 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567690 4967 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567700 4967 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567709 4967 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567717 4967 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567727 4967 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567736 4967 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567745 4967 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567754 4967 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567763 4967 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567772 4967 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567782 4967 feature_gate.go:330] unrecognized feature gate: Example Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567790 4967 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567798 4967 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567807 4967 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567815 4967 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567824 4967 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567832 4967 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567858 4967 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567867 4967 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567875 4967 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567884 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567892 4967 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567900 4967 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567908 4967 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567917 4967 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567925 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567934 4967 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567942 4967 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567950 4967 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567958 4967 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567969 4967 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567980 4967 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567989 4967 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.567998 4967 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.568006 4967 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.568014 4967 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.568023 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.568031 4967 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.568041 4967 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.568053 4967 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.568062 4967 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.568077 4967 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.568086 4967 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.568095 4967 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.568103 4967 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.568111 4967 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.568147 4967 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.568156 4967 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.568164 4967 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.568174 4967 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.568183 4967 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.568194 4967 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.568207 4967 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.568215 4967 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.568224 4967 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.568232 4967 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.568240 4967 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.568249 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.568257 4967 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.568265 4967 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.568274 4967 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568434 4967 flags.go:64] FLAG: --address="0.0.0.0" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568452 4967 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568468 4967 flags.go:64] FLAG: --anonymous-auth="true" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568481 4967 flags.go:64] FLAG: --application-metrics-count-limit="100" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568493 4967 flags.go:64] FLAG: --authentication-token-webhook="false" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568502 4967 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568515 4967 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568528 4967 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568538 4967 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568548 4967 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568558 4967 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568569 4967 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568578 4967 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568588 4967 flags.go:64] FLAG: --cgroup-root="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568597 4967 flags.go:64] FLAG: --cgroups-per-qos="true" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568607 4967 flags.go:64] FLAG: --client-ca-file="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568617 4967 flags.go:64] FLAG: --cloud-config="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568626 4967 flags.go:64] FLAG: --cloud-provider="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568635 4967 flags.go:64] FLAG: --cluster-dns="[]" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568646 4967 flags.go:64] FLAG: --cluster-domain="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568655 4967 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568665 4967 flags.go:64] FLAG: --config-dir="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568675 4967 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568685 4967 flags.go:64] FLAG: --container-log-max-files="5" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568696 4967 flags.go:64] FLAG: --container-log-max-size="10Mi" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568706 4967 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568715 4967 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568726 4967 flags.go:64] FLAG: --containerd-namespace="k8s.io" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568736 4967 flags.go:64] FLAG: --contention-profiling="false" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568747 4967 flags.go:64] FLAG: --cpu-cfs-quota="true" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568757 4967 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568767 4967 flags.go:64] FLAG: --cpu-manager-policy="none" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568777 4967 flags.go:64] FLAG: --cpu-manager-policy-options="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568789 4967 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568798 4967 flags.go:64] FLAG: --enable-controller-attach-detach="true" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568809 4967 flags.go:64] FLAG: --enable-debugging-handlers="true" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568818 4967 flags.go:64] FLAG: --enable-load-reader="false" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568829 4967 flags.go:64] FLAG: --enable-server="true" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568838 4967 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568850 4967 flags.go:64] FLAG: --event-burst="100" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568861 4967 flags.go:64] FLAG: --event-qps="50" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568870 4967 flags.go:64] FLAG: --event-storage-age-limit="default=0" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568880 4967 flags.go:64] FLAG: --event-storage-event-limit="default=0" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568893 4967 flags.go:64] FLAG: --eviction-hard="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568904 4967 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568913 4967 flags.go:64] FLAG: --eviction-minimum-reclaim="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568923 4967 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568936 4967 flags.go:64] FLAG: --eviction-soft="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568946 4967 flags.go:64] FLAG: --eviction-soft-grace-period="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568956 4967 flags.go:64] FLAG: --exit-on-lock-contention="false" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568965 4967 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568975 4967 flags.go:64] FLAG: --experimental-mounter-path="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568985 4967 flags.go:64] FLAG: --fail-cgroupv1="false" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.568996 4967 flags.go:64] FLAG: --fail-swap-on="true" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569005 4967 flags.go:64] FLAG: --feature-gates="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569016 4967 flags.go:64] FLAG: --file-check-frequency="20s" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569026 4967 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569036 4967 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569046 4967 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569056 4967 flags.go:64] FLAG: --healthz-port="10248" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569066 4967 flags.go:64] FLAG: --help="false" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569085 4967 flags.go:64] FLAG: --hostname-override="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569096 4967 flags.go:64] FLAG: --housekeeping-interval="10s" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569108 4967 flags.go:64] FLAG: --http-check-frequency="20s" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569151 4967 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569162 4967 flags.go:64] FLAG: --image-credential-provider-config="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569173 4967 flags.go:64] FLAG: --image-gc-high-threshold="85" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569184 4967 flags.go:64] FLAG: --image-gc-low-threshold="80" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569196 4967 flags.go:64] FLAG: --image-service-endpoint="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569207 4967 flags.go:64] FLAG: --kernel-memcg-notification="false" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569219 4967 flags.go:64] FLAG: --kube-api-burst="100" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569230 4967 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569240 4967 flags.go:64] FLAG: --kube-api-qps="50" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569252 4967 flags.go:64] FLAG: --kube-reserved="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569265 4967 flags.go:64] FLAG: --kube-reserved-cgroup="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569275 4967 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569286 4967 flags.go:64] FLAG: --kubelet-cgroups="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569296 4967 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569307 4967 flags.go:64] FLAG: --lock-file="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569322 4967 flags.go:64] FLAG: --log-cadvisor-usage="false" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569334 4967 flags.go:64] FLAG: --log-flush-frequency="5s" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569346 4967 flags.go:64] FLAG: --log-json-info-buffer-size="0" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569363 4967 flags.go:64] FLAG: --log-json-split-stream="false" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569373 4967 flags.go:64] FLAG: --log-text-info-buffer-size="0" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569385 4967 flags.go:64] FLAG: --log-text-split-stream="false" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569396 4967 flags.go:64] FLAG: --logging-format="text" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569407 4967 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569419 4967 flags.go:64] FLAG: --make-iptables-util-chains="true" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569429 4967 flags.go:64] FLAG: --manifest-url="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569440 4967 flags.go:64] FLAG: --manifest-url-header="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569454 4967 flags.go:64] FLAG: --max-housekeeping-interval="15s" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569465 4967 flags.go:64] FLAG: --max-open-files="1000000" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569478 4967 flags.go:64] FLAG: --max-pods="110" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569489 4967 flags.go:64] FLAG: --maximum-dead-containers="-1" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569500 4967 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569511 4967 flags.go:64] FLAG: --memory-manager-policy="None" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569521 4967 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569532 4967 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569544 4967 flags.go:64] FLAG: --node-ip="192.168.126.11" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569556 4967 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569582 4967 flags.go:64] FLAG: --node-status-max-images="50" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569592 4967 flags.go:64] FLAG: --node-status-update-frequency="10s" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569603 4967 flags.go:64] FLAG: --oom-score-adj="-999" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569614 4967 flags.go:64] FLAG: --pod-cidr="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569624 4967 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569640 4967 flags.go:64] FLAG: --pod-manifest-path="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569652 4967 flags.go:64] FLAG: --pod-max-pids="-1" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569663 4967 flags.go:64] FLAG: --pods-per-core="0" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569673 4967 flags.go:64] FLAG: --port="10250" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569684 4967 flags.go:64] FLAG: --protect-kernel-defaults="false" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569694 4967 flags.go:64] FLAG: --provider-id="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569705 4967 flags.go:64] FLAG: --qos-reserved="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569715 4967 flags.go:64] FLAG: --read-only-port="10255" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569726 4967 flags.go:64] FLAG: --register-node="true" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569737 4967 flags.go:64] FLAG: --register-schedulable="true" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569751 4967 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569768 4967 flags.go:64] FLAG: --registry-burst="10" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569779 4967 flags.go:64] FLAG: --registry-qps="5" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569789 4967 flags.go:64] FLAG: --reserved-cpus="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569799 4967 flags.go:64] FLAG: --reserved-memory="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569813 4967 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569824 4967 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569835 4967 flags.go:64] FLAG: --rotate-certificates="false" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569846 4967 flags.go:64] FLAG: --rotate-server-certificates="false" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569857 4967 flags.go:64] FLAG: --runonce="false" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569868 4967 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569880 4967 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569891 4967 flags.go:64] FLAG: --seccomp-default="false" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569902 4967 flags.go:64] FLAG: --serialize-image-pulls="true" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569913 4967 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569923 4967 flags.go:64] FLAG: --storage-driver-db="cadvisor" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569936 4967 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569947 4967 flags.go:64] FLAG: --storage-driver-password="root" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569958 4967 flags.go:64] FLAG: --storage-driver-secure="false" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569968 4967 flags.go:64] FLAG: --storage-driver-table="stats" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569978 4967 flags.go:64] FLAG: --storage-driver-user="root" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.569990 4967 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.570000 4967 flags.go:64] FLAG: --sync-frequency="1m0s" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.570012 4967 flags.go:64] FLAG: --system-cgroups="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.570022 4967 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.570041 4967 flags.go:64] FLAG: --system-reserved-cgroup="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.570052 4967 flags.go:64] FLAG: --tls-cert-file="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.570062 4967 flags.go:64] FLAG: --tls-cipher-suites="[]" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.570084 4967 flags.go:64] FLAG: --tls-min-version="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.570094 4967 flags.go:64] FLAG: --tls-private-key-file="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.570103 4967 flags.go:64] FLAG: --topology-manager-policy="none" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.570113 4967 flags.go:64] FLAG: --topology-manager-policy-options="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.570150 4967 flags.go:64] FLAG: --topology-manager-scope="container" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.570160 4967 flags.go:64] FLAG: --v="2" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.570172 4967 flags.go:64] FLAG: --version="false" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.570184 4967 flags.go:64] FLAG: --vmodule="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.570197 4967 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.570209 4967 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570423 4967 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570433 4967 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570443 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570451 4967 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570461 4967 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570472 4967 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570483 4967 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570493 4967 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570508 4967 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570517 4967 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570528 4967 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570538 4967 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570548 4967 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570557 4967 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570567 4967 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570576 4967 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570585 4967 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570593 4967 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570602 4967 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570610 4967 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570619 4967 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570627 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570635 4967 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570644 4967 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570658 4967 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570667 4967 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570699 4967 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570708 4967 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570716 4967 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570725 4967 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570734 4967 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570742 4967 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570750 4967 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570758 4967 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570768 4967 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570777 4967 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570785 4967 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570794 4967 feature_gate.go:330] unrecognized feature gate: Example Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570802 4967 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570810 4967 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570824 4967 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570833 4967 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570842 4967 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570850 4967 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570859 4967 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570868 4967 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570877 4967 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570885 4967 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570893 4967 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570902 4967 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570913 4967 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570925 4967 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570935 4967 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570944 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570954 4967 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570963 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570972 4967 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570980 4967 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570988 4967 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.570997 4967 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.571005 4967 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.571013 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.571022 4967 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.571030 4967 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.571038 4967 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.571046 4967 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.571054 4967 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.571063 4967 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.571072 4967 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.571080 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.571092 4967 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.571152 4967 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.583199 4967 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.583247 4967 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583381 4967 feature_gate.go:330] unrecognized feature gate: Example Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583392 4967 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583400 4967 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583406 4967 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583412 4967 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583418 4967 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583425 4967 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583430 4967 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583436 4967 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583442 4967 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583448 4967 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583454 4967 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583459 4967 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583467 4967 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583475 4967 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583481 4967 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583487 4967 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583494 4967 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583501 4967 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583509 4967 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583515 4967 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583521 4967 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583527 4967 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583534 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583540 4967 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583547 4967 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583554 4967 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583560 4967 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583566 4967 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583571 4967 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583576 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583581 4967 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583586 4967 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583590 4967 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583594 4967 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583598 4967 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583602 4967 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583606 4967 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583610 4967 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583614 4967 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583618 4967 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583622 4967 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583628 4967 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583633 4967 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583638 4967 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583642 4967 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583647 4967 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583651 4967 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583656 4967 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583660 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583664 4967 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583668 4967 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583672 4967 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583677 4967 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583681 4967 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583686 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583690 4967 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583694 4967 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583698 4967 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583702 4967 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583706 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583711 4967 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583715 4967 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583720 4967 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583725 4967 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583729 4967 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583733 4967 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583737 4967 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583742 4967 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583748 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583752 4967 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.583759 4967 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583888 4967 feature_gate.go:330] unrecognized feature gate: InsightsConfig Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583896 4967 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583902 4967 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583908 4967 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583912 4967 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583917 4967 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583921 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583925 4967 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583930 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583934 4967 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583952 4967 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583957 4967 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583961 4967 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583965 4967 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583969 4967 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583973 4967 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583978 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583984 4967 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.583989 4967 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584006 4967 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584011 4967 feature_gate.go:330] unrecognized feature gate: Example Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584015 4967 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584019 4967 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584023 4967 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584029 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584034 4967 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584039 4967 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584043 4967 feature_gate.go:330] unrecognized feature gate: SignatureStores Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584047 4967 feature_gate.go:330] unrecognized feature gate: OVNObservability Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584051 4967 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584055 4967 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584059 4967 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584063 4967 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584067 4967 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584071 4967 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584075 4967 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584083 4967 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584087 4967 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584091 4967 feature_gate.go:330] unrecognized feature gate: NewOLM Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584095 4967 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584099 4967 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584103 4967 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584107 4967 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584111 4967 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584117 4967 feature_gate.go:330] unrecognized feature gate: PinnedImages Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584134 4967 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584138 4967 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584143 4967 feature_gate.go:330] unrecognized feature gate: PlatformOperators Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584147 4967 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584151 4967 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584156 4967 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584161 4967 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584167 4967 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584173 4967 feature_gate.go:330] unrecognized feature gate: GatewayAPI Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584178 4967 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584183 4967 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584188 4967 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584192 4967 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584196 4967 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584200 4967 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584204 4967 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584209 4967 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584213 4967 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584218 4967 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584222 4967 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584226 4967 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584230 4967 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584233 4967 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584238 4967 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584242 4967 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.584245 4967 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.584252 4967 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.585119 4967 server.go:940] "Client rotation is on, will bootstrap in background" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.589215 4967 bootstrap.go:85] "Current kubeconfig file contents are still valid, no bootstrap necessary" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.589296 4967 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.590764 4967 server.go:997] "Starting client certificate rotation" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.590788 4967 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.591013 4967 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2026-02-24 05:52:08 +0000 UTC, rotation deadline is 2026-01-06 02:01:27.911417669 +0000 UTC Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.591175 4967 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 2086h10m15.320248578s for next certificate rotation Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.621836 4967 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.627277 4967 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.650110 4967 log.go:25] "Validated CRI v1 runtime API" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.687155 4967 log.go:25] "Validated CRI v1 image API" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.689014 4967 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.696141 4967 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2025-10-11-03-47-08-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.696180 4967 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.720514 4967 manager.go:217] Machine: {Timestamp:2025-10-11 03:51:12.717294359 +0000 UTC m=+0.680503332 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654128640 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:39750116-adfb-459d-8c17-e38159f0ae3a BootID:2f687e2b-93fa-4909-96cb-dd8768ce3482 Filesystems:[{Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true} {Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730825728 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827064320 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:bc:ab:c4 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:bc:ab:c4 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:d9:f2:4f Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:74:60:2e Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:7a:8b:1c Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:c6:19:55 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:2a:4d:ce:84:a7:f9 Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:8a:92:01:b7:e1:75 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654128640 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.720804 4967 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.720973 4967 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.723239 4967 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.723489 4967 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.723542 4967 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.723773 4967 topology_manager.go:138] "Creating topology manager with none policy" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.723784 4967 container_manager_linux.go:303] "Creating device plugin manager" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.724267 4967 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.724321 4967 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.724533 4967 state_mem.go:36] "Initialized new in-memory state store" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.724628 4967 server.go:1245] "Using root directory" path="/var/lib/kubelet" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.728648 4967 kubelet.go:418] "Attempting to sync node with API server" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.728778 4967 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.728894 4967 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.728920 4967 kubelet.go:324] "Adding apiserver pod source" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.728941 4967 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.734335 4967 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.735694 4967 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.736322 4967 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Oct 11 03:51:12 crc kubenswrapper[4967]: E1011 03:51:12.736436 4967 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.736440 4967 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Oct 11 03:51:12 crc kubenswrapper[4967]: E1011 03:51:12.736551 4967 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.737886 4967 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.741655 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.741700 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.741715 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.741732 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.741754 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.741767 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.741782 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.741803 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.741818 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.741831 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.741871 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.741885 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.744558 4967 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.749144 4967 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.749363 4967 server.go:1280] "Started kubelet" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.750670 4967 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.750669 4967 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.751677 4967 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Oct 11 03:51:12 crc systemd[1]: Started Kubernetes Kubelet. Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.752622 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.753244 4967 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.753840 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2026-02-24 05:53:03 +0000 UTC, rotation deadline is 2025-12-17 17:59:53.715119987 +0000 UTC Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.753904 4967 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 1622h8m40.961220446s for next certificate rotation Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.753972 4967 volume_manager.go:287] "The desired_state_of_world populator starts" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.753991 4967 volume_manager.go:289] "Starting Kubelet Volume Manager" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.754035 4967 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Oct 11 03:51:12 crc kubenswrapper[4967]: E1011 03:51:12.754191 4967 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.754694 4967 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Oct 11 03:51:12 crc kubenswrapper[4967]: E1011 03:51:12.754786 4967 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.754908 4967 server.go:460] "Adding debug handlers to kubelet server" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.755265 4967 factory.go:55] Registering systemd factory Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.755656 4967 factory.go:221] Registration of the systemd container factory successfully Oct 11 03:51:12 crc kubenswrapper[4967]: E1011 03:51:12.755680 4967 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="200ms" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.757283 4967 factory.go:153] Registering CRI-O factory Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.757319 4967 factory.go:221] Registration of the crio container factory successfully Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.757393 4967 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.757422 4967 factory.go:103] Registering Raw factory Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.757444 4967 manager.go:1196] Started watching for new ooms in manager Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.759039 4967 manager.go:319] Starting recovery of all containers Oct 11 03:51:12 crc kubenswrapper[4967]: E1011 03:51:12.759579 4967 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.50:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186d535962f78e6f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-11 03:51:12.749317743 +0000 UTC m=+0.712526706,LastTimestamp:2025-10-11 03:51:12.749317743 +0000 UTC m=+0.712526706,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.775998 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.776605 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.776862 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.777517 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.777576 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.777611 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.777661 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.777690 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.777730 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.777762 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.777855 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.777881 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.777901 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.777931 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.777957 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.777979 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.778043 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.778112 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.778232 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.778291 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.778322 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.778361 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.778390 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.778432 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.778481 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.778511 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.778554 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.778600 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.778630 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.778666 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.778691 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.778720 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.778753 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.778782 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.778815 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.778865 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.778891 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.778939 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.779183 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.779287 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.779315 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.779344 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.779363 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.779389 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.779408 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.779426 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.779453 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.779474 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.779493 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.779519 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.779538 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.779562 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.779605 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.779643 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.779674 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.779697 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.779723 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.779742 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.779770 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.779790 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.779814 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.779834 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.779852 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.779895 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.779915 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.779936 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.779954 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.779971 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.779995 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780012 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780031 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780085 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780109 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780155 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780174 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780193 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780217 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780235 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780261 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780280 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780298 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780323 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780344 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780367 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780389 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780406 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780429 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780450 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780477 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780521 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780541 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780563 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780582 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780614 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780659 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780677 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780714 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780732 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780748 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780773 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780792 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780814 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780832 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780853 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780887 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780919 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780950 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.780973 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781002 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781025 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781055 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781083 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781113 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781196 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781215 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781238 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781261 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781287 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781305 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781324 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781413 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781432 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781456 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781474 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781494 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781517 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781537 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781575 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781599 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781648 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781679 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781698 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781722 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781759 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781779 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781817 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781836 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781861 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781879 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781895 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.781918 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.782667 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.782692 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.782708 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.784550 4967 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.784581 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.784598 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.784613 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.784626 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.784642 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.784655 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.784669 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.784682 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.784694 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.784707 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.784720 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.784807 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.784823 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.784876 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.784892 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.784905 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.784923 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.784937 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.784949 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.784962 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.784976 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.784988 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785000 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785011 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785026 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785061 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785080 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785094 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785111 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785208 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785223 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785236 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785251 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785266 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785279 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785292 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785307 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785321 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785334 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785348 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785367 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785382 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785396 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785411 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785425 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785440 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785458 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785470 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785484 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785497 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785510 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785522 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785537 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785550 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785562 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785575 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785588 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785603 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785616 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785630 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785643 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785656 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785671 4967 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785683 4967 reconstruct.go:97] "Volume reconstruction finished" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.785691 4967 reconciler.go:26] "Reconciler: start to sync state" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.789207 4967 manager.go:324] Recovery completed Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.797894 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.800049 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.800090 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.800101 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.802874 4967 cpu_manager.go:225] "Starting CPU manager" policy="none" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.802895 4967 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.802914 4967 state_mem.go:36] "Initialized new in-memory state store" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.811952 4967 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.813441 4967 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.813541 4967 status_manager.go:217] "Starting to sync pod status with apiserver" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.813574 4967 kubelet.go:2335] "Starting kubelet main sync loop" Oct 11 03:51:12 crc kubenswrapper[4967]: E1011 03:51:12.813675 4967 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Oct 11 03:51:12 crc kubenswrapper[4967]: W1011 03:51:12.814581 4967 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Oct 11 03:51:12 crc kubenswrapper[4967]: E1011 03:51:12.814746 4967 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.819862 4967 policy_none.go:49] "None policy: Start" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.820715 4967 memory_manager.go:170] "Starting memorymanager" policy="None" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.820833 4967 state_mem.go:35] "Initializing new in-memory state store" Oct 11 03:51:12 crc kubenswrapper[4967]: E1011 03:51:12.856488 4967 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.868171 4967 manager.go:334] "Starting Device Plugin manager" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.868299 4967 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.868319 4967 server.go:79] "Starting device plugin registration server" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.868769 4967 eviction_manager.go:189] "Eviction manager: starting control loop" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.868787 4967 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.869430 4967 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.869529 4967 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.869540 4967 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Oct 11 03:51:12 crc kubenswrapper[4967]: E1011 03:51:12.875635 4967 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.914389 4967 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc","openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.914571 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.917293 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.917364 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.917383 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.917638 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.918276 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.918376 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.919780 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.919852 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.919876 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.919960 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.919998 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.920017 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.920245 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.920464 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.920529 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.921448 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.921487 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.921503 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.921645 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.921818 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.921876 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.921897 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.921822 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.921980 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.923201 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.923250 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.923270 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.923636 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.923694 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.923711 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.923904 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.924182 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.924315 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.925176 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.925220 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.925241 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.925529 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.925575 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.925685 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.925713 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.925731 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.926637 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.926675 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.926691 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:12 crc kubenswrapper[4967]: E1011 03:51:12.957423 4967 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="400ms" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.969156 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.970321 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.970362 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.970379 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.970411 4967 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 11 03:51:12 crc kubenswrapper[4967]: E1011 03:51:12.971482 4967 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.50:6443: connect: connection refused" node="crc" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.988813 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.989367 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.989490 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.989574 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.991884 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.992236 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.992586 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.992675 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.993894 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.994847 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.994898 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.994933 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.994963 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.994995 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 11 03:51:12 crc kubenswrapper[4967]: I1011 03:51:12.995028 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:51:12 crc kubenswrapper[4967]: E1011 03:51:12.996719 4967 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.50:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.186d535962f78e6f default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2025-10-11 03:51:12.749317743 +0000 UTC m=+0.712526706,LastTimestamp:2025-10-11 03:51:12.749317743 +0000 UTC m=+0.712526706,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.096924 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.097499 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.097316 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.097677 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.097987 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.098158 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.098243 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.098266 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.098336 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.098358 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.098376 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.098342 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.098470 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.098538 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.098568 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.098605 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.098637 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.098575 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.098651 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.098670 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.098777 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.098772 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.098741 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.098834 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.098866 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.098899 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.098930 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.098932 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.099028 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.099138 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.172005 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.173874 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.173964 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.173989 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.174032 4967 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 11 03:51:13 crc kubenswrapper[4967]: E1011 03:51:13.175849 4967 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.50:6443: connect: connection refused" node="crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.251468 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.268186 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.287343 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.296359 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.304371 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Oct 11 03:51:13 crc kubenswrapper[4967]: W1011 03:51:13.305595 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-c3c4dfdb127d186c61d9e940e99e060172a286ae7cacf55c7586e547fc6f8b1e WatchSource:0}: Error finding container c3c4dfdb127d186c61d9e940e99e060172a286ae7cacf55c7586e547fc6f8b1e: Status 404 returned error can't find the container with id c3c4dfdb127d186c61d9e940e99e060172a286ae7cacf55c7586e547fc6f8b1e Oct 11 03:51:13 crc kubenswrapper[4967]: W1011 03:51:13.311945 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-437c9071b6e1951dacdeeca1de3a48c60c22f4fae957c28c4858208338eb49aa WatchSource:0}: Error finding container 437c9071b6e1951dacdeeca1de3a48c60c22f4fae957c28c4858208338eb49aa: Status 404 returned error can't find the container with id 437c9071b6e1951dacdeeca1de3a48c60c22f4fae957c28c4858208338eb49aa Oct 11 03:51:13 crc kubenswrapper[4967]: W1011 03:51:13.320019 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-a7739080caaa320fa7ecef66a7cdb6afd7a6c8fb0d898beb03b9e266c24e7b2d WatchSource:0}: Error finding container a7739080caaa320fa7ecef66a7cdb6afd7a6c8fb0d898beb03b9e266c24e7b2d: Status 404 returned error can't find the container with id a7739080caaa320fa7ecef66a7cdb6afd7a6c8fb0d898beb03b9e266c24e7b2d Oct 11 03:51:13 crc kubenswrapper[4967]: W1011 03:51:13.324830 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-453f7be9fd41c9826de69553518bcacaee060d60a1536c56c55899f15e95ff51 WatchSource:0}: Error finding container 453f7be9fd41c9826de69553518bcacaee060d60a1536c56c55899f15e95ff51: Status 404 returned error can't find the container with id 453f7be9fd41c9826de69553518bcacaee060d60a1536c56c55899f15e95ff51 Oct 11 03:51:13 crc kubenswrapper[4967]: W1011 03:51:13.330889 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-8447e5e7a0b8f9b006d6600766947f3a8edbabeec226bf504966b1bdc72586fd WatchSource:0}: Error finding container 8447e5e7a0b8f9b006d6600766947f3a8edbabeec226bf504966b1bdc72586fd: Status 404 returned error can't find the container with id 8447e5e7a0b8f9b006d6600766947f3a8edbabeec226bf504966b1bdc72586fd Oct 11 03:51:13 crc kubenswrapper[4967]: E1011 03:51:13.358693 4967 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="800ms" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.576904 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.579097 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.579155 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.579167 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.579190 4967 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 11 03:51:13 crc kubenswrapper[4967]: E1011 03:51:13.579685 4967 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.50:6443: connect: connection refused" node="crc" Oct 11 03:51:13 crc kubenswrapper[4967]: W1011 03:51:13.624491 4967 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Oct 11 03:51:13 crc kubenswrapper[4967]: E1011 03:51:13.624592 4967 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Oct 11 03:51:13 crc kubenswrapper[4967]: W1011 03:51:13.728921 4967 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Oct 11 03:51:13 crc kubenswrapper[4967]: E1011 03:51:13.728996 4967 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.750285 4967 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Oct 11 03:51:13 crc kubenswrapper[4967]: W1011 03:51:13.770532 4967 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Oct 11 03:51:13 crc kubenswrapper[4967]: E1011 03:51:13.770677 4967 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.818152 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"453f7be9fd41c9826de69553518bcacaee060d60a1536c56c55899f15e95ff51"} Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.819840 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"a7739080caaa320fa7ecef66a7cdb6afd7a6c8fb0d898beb03b9e266c24e7b2d"} Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.821101 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c3c4dfdb127d186c61d9e940e99e060172a286ae7cacf55c7586e547fc6f8b1e"} Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.823815 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"437c9071b6e1951dacdeeca1de3a48c60c22f4fae957c28c4858208338eb49aa"} Oct 11 03:51:13 crc kubenswrapper[4967]: I1011 03:51:13.824846 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"8447e5e7a0b8f9b006d6600766947f3a8edbabeec226bf504966b1bdc72586fd"} Oct 11 03:51:14 crc kubenswrapper[4967]: E1011 03:51:14.160154 4967 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="1.6s" Oct 11 03:51:14 crc kubenswrapper[4967]: W1011 03:51:14.276747 4967 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Oct 11 03:51:14 crc kubenswrapper[4967]: E1011 03:51:14.277188 4967 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.380167 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.381669 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.381895 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.381913 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.381963 4967 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 11 03:51:14 crc kubenswrapper[4967]: E1011 03:51:14.382552 4967 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.50:6443: connect: connection refused" node="crc" Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.750472 4967 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.832508 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc"} Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.832561 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344"} Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.832575 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d"} Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.832589 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c"} Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.832615 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.834075 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.834131 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.834142 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.835125 4967 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39" exitCode=0 Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.835188 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.835182 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39"} Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.836854 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.836893 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.836908 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.839638 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.839704 4967 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5" exitCode=0 Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.839953 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.840020 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5"} Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.841375 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.841418 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.841449 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.841467 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.841425 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.841526 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.843178 4967 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="e84c5697ab006ad26c4ebc739bf09997d679c605f84ffe355d30f2b32f4f0e54" exitCode=0 Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.843266 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"e84c5697ab006ad26c4ebc739bf09997d679c605f84ffe355d30f2b32f4f0e54"} Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.843337 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.845440 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.845472 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.845483 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.846102 4967 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="df874a11e199cea6058abe0e5bb63faf7246199fa02093f05720bc41dbb7a81a" exitCode=0 Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.846189 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"df874a11e199cea6058abe0e5bb63faf7246199fa02093f05720bc41dbb7a81a"} Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.846193 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.847683 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.847707 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:14 crc kubenswrapper[4967]: I1011 03:51:14.847717 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.422805 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 11 03:51:15 crc kubenswrapper[4967]: W1011 03:51:15.495202 4967 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Oct 11 03:51:15 crc kubenswrapper[4967]: E1011 03:51:15.495298 4967 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.50:6443: connect: connection refused" logger="UnhandledError" Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.750368 4967 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.50:6443: connect: connection refused Oct 11 03:51:15 crc kubenswrapper[4967]: E1011 03:51:15.761040 4967 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.50:6443: connect: connection refused" interval="3.2s" Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.852806 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"d3f23f4ffaeac5c8606def100262484a45cd86296beb91735215c57955e45067"} Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.852872 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"fef6a4dbe93461c6352741acfc9439a22668f46dd725fcf80fb37d93c88e9d11"} Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.852894 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"b2b90baa6dbb24ee4c2deea587bada1820917fd963caaf91210c25409b254c60"} Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.852899 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.854273 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.854305 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.854316 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.858801 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181"} Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.858831 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f"} Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.858845 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e"} Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.858858 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f"} Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.860728 4967 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f" exitCode=0 Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.860894 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.861034 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f"} Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.861757 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.861812 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.861823 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.864269 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.864282 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.864551 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"7fc96d2d9f48b01b4c4f1f4bc3e058427ffc11b3b73c0ad8a89dbd690a098b67"} Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.865754 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.865779 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.865789 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.866358 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.866378 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.866388 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.982971 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.984514 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.984548 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.984556 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:15 crc kubenswrapper[4967]: I1011 03:51:15.984578 4967 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 11 03:51:15 crc kubenswrapper[4967]: E1011 03:51:15.985045 4967 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.50:6443: connect: connection refused" node="crc" Oct 11 03:51:16 crc kubenswrapper[4967]: I1011 03:51:16.871166 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42"} Oct 11 03:51:16 crc kubenswrapper[4967]: I1011 03:51:16.871971 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:16 crc kubenswrapper[4967]: I1011 03:51:16.873404 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:16 crc kubenswrapper[4967]: I1011 03:51:16.873462 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:16 crc kubenswrapper[4967]: I1011 03:51:16.873486 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:16 crc kubenswrapper[4967]: I1011 03:51:16.874877 4967 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb" exitCode=0 Oct 11 03:51:16 crc kubenswrapper[4967]: I1011 03:51:16.875000 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:16 crc kubenswrapper[4967]: I1011 03:51:16.875018 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:16 crc kubenswrapper[4967]: I1011 03:51:16.875105 4967 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 11 03:51:16 crc kubenswrapper[4967]: I1011 03:51:16.875174 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:16 crc kubenswrapper[4967]: I1011 03:51:16.875110 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:16 crc kubenswrapper[4967]: I1011 03:51:16.875106 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb"} Oct 11 03:51:16 crc kubenswrapper[4967]: I1011 03:51:16.877149 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:16 crc kubenswrapper[4967]: I1011 03:51:16.877167 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:16 crc kubenswrapper[4967]: I1011 03:51:16.877210 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:16 crc kubenswrapper[4967]: I1011 03:51:16.877152 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:16 crc kubenswrapper[4967]: I1011 03:51:16.877237 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:16 crc kubenswrapper[4967]: I1011 03:51:16.877260 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:16 crc kubenswrapper[4967]: I1011 03:51:16.877351 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:16 crc kubenswrapper[4967]: I1011 03:51:16.877213 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:16 crc kubenswrapper[4967]: I1011 03:51:16.877412 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:16 crc kubenswrapper[4967]: I1011 03:51:16.877740 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:16 crc kubenswrapper[4967]: I1011 03:51:16.877784 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:16 crc kubenswrapper[4967]: I1011 03:51:16.877806 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:17 crc kubenswrapper[4967]: I1011 03:51:17.881246 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636"} Oct 11 03:51:17 crc kubenswrapper[4967]: I1011 03:51:17.881313 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69"} Oct 11 03:51:17 crc kubenswrapper[4967]: I1011 03:51:17.881339 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d"} Oct 11 03:51:17 crc kubenswrapper[4967]: I1011 03:51:17.881359 4967 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 11 03:51:17 crc kubenswrapper[4967]: I1011 03:51:17.881441 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:17 crc kubenswrapper[4967]: I1011 03:51:17.884347 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:17 crc kubenswrapper[4967]: I1011 03:51:17.884388 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:17 crc kubenswrapper[4967]: I1011 03:51:17.884399 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:18 crc kubenswrapper[4967]: I1011 03:51:18.893132 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8"} Oct 11 03:51:18 crc kubenswrapper[4967]: I1011 03:51:18.893216 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1"} Oct 11 03:51:18 crc kubenswrapper[4967]: I1011 03:51:18.893309 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:18 crc kubenswrapper[4967]: I1011 03:51:18.894418 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:18 crc kubenswrapper[4967]: I1011 03:51:18.894484 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:18 crc kubenswrapper[4967]: I1011 03:51:18.894508 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:18 crc kubenswrapper[4967]: I1011 03:51:18.973892 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 11 03:51:18 crc kubenswrapper[4967]: I1011 03:51:18.974144 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:18 crc kubenswrapper[4967]: I1011 03:51:18.975573 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:18 crc kubenswrapper[4967]: I1011 03:51:18.975638 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:18 crc kubenswrapper[4967]: I1011 03:51:18.975666 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:19 crc kubenswrapper[4967]: I1011 03:51:19.186012 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:19 crc kubenswrapper[4967]: I1011 03:51:19.187647 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:19 crc kubenswrapper[4967]: I1011 03:51:19.187720 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:19 crc kubenswrapper[4967]: I1011 03:51:19.187745 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:19 crc kubenswrapper[4967]: I1011 03:51:19.187786 4967 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 11 03:51:19 crc kubenswrapper[4967]: I1011 03:51:19.390739 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:51:19 crc kubenswrapper[4967]: I1011 03:51:19.390966 4967 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 11 03:51:19 crc kubenswrapper[4967]: I1011 03:51:19.391021 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:19 crc kubenswrapper[4967]: I1011 03:51:19.392525 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:19 crc kubenswrapper[4967]: I1011 03:51:19.392602 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:19 crc kubenswrapper[4967]: I1011 03:51:19.392628 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:19 crc kubenswrapper[4967]: I1011 03:51:19.896807 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:19 crc kubenswrapper[4967]: I1011 03:51:19.898132 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:19 crc kubenswrapper[4967]: I1011 03:51:19.898187 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:19 crc kubenswrapper[4967]: I1011 03:51:19.898203 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:20 crc kubenswrapper[4967]: I1011 03:51:20.791689 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Oct 11 03:51:20 crc kubenswrapper[4967]: I1011 03:51:20.899928 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:20 crc kubenswrapper[4967]: I1011 03:51:20.901375 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:20 crc kubenswrapper[4967]: I1011 03:51:20.901493 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:20 crc kubenswrapper[4967]: I1011 03:51:20.901513 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:20 crc kubenswrapper[4967]: I1011 03:51:20.943727 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:51:20 crc kubenswrapper[4967]: I1011 03:51:20.943968 4967 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 11 03:51:20 crc kubenswrapper[4967]: I1011 03:51:20.944029 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:20 crc kubenswrapper[4967]: I1011 03:51:20.945859 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:20 crc kubenswrapper[4967]: I1011 03:51:20.945922 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:20 crc kubenswrapper[4967]: I1011 03:51:20.945938 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:21 crc kubenswrapper[4967]: I1011 03:51:21.248687 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:51:21 crc kubenswrapper[4967]: I1011 03:51:21.903304 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:21 crc kubenswrapper[4967]: I1011 03:51:21.904583 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:21 crc kubenswrapper[4967]: I1011 03:51:21.904672 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:21 crc kubenswrapper[4967]: I1011 03:51:21.904693 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:22 crc kubenswrapper[4967]: I1011 03:51:22.616985 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 11 03:51:22 crc kubenswrapper[4967]: I1011 03:51:22.617330 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:22 crc kubenswrapper[4967]: I1011 03:51:22.619044 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:22 crc kubenswrapper[4967]: I1011 03:51:22.619189 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:22 crc kubenswrapper[4967]: I1011 03:51:22.619222 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:22 crc kubenswrapper[4967]: E1011 03:51:22.875825 4967 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Oct 11 03:51:23 crc kubenswrapper[4967]: I1011 03:51:23.004055 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Oct 11 03:51:23 crc kubenswrapper[4967]: I1011 03:51:23.004407 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:23 crc kubenswrapper[4967]: I1011 03:51:23.005994 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:23 crc kubenswrapper[4967]: I1011 03:51:23.006047 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:23 crc kubenswrapper[4967]: I1011 03:51:23.006063 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:23 crc kubenswrapper[4967]: I1011 03:51:23.658585 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 11 03:51:23 crc kubenswrapper[4967]: I1011 03:51:23.658807 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:23 crc kubenswrapper[4967]: I1011 03:51:23.660365 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:23 crc kubenswrapper[4967]: I1011 03:51:23.660420 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:23 crc kubenswrapper[4967]: I1011 03:51:23.660439 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:24 crc kubenswrapper[4967]: I1011 03:51:24.038301 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 11 03:51:24 crc kubenswrapper[4967]: I1011 03:51:24.038516 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:24 crc kubenswrapper[4967]: I1011 03:51:24.040341 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:24 crc kubenswrapper[4967]: I1011 03:51:24.040411 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:24 crc kubenswrapper[4967]: I1011 03:51:24.040425 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:24 crc kubenswrapper[4967]: I1011 03:51:24.044635 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 11 03:51:24 crc kubenswrapper[4967]: I1011 03:51:24.912002 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:24 crc kubenswrapper[4967]: I1011 03:51:24.913002 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:24 crc kubenswrapper[4967]: I1011 03:51:24.913043 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:24 crc kubenswrapper[4967]: I1011 03:51:24.913056 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:24 crc kubenswrapper[4967]: I1011 03:51:24.919433 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 11 03:51:25 crc kubenswrapper[4967]: I1011 03:51:25.914475 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:25 crc kubenswrapper[4967]: I1011 03:51:25.915812 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:25 crc kubenswrapper[4967]: I1011 03:51:25.915861 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:25 crc kubenswrapper[4967]: I1011 03:51:25.915873 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:26 crc kubenswrapper[4967]: I1011 03:51:26.151937 4967 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Liveness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 11 03:51:26 crc kubenswrapper[4967]: I1011 03:51:26.152134 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 11 03:51:26 crc kubenswrapper[4967]: W1011 03:51:26.472512 4967 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 11 03:51:26 crc kubenswrapper[4967]: I1011 03:51:26.472642 4967 trace.go:236] Trace[1577961454]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (11-Oct-2025 03:51:16.471) (total time: 10001ms): Oct 11 03:51:26 crc kubenswrapper[4967]: Trace[1577961454]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (03:51:26.472) Oct 11 03:51:26 crc kubenswrapper[4967]: Trace[1577961454]: [10.00119842s] [10.00119842s] END Oct 11 03:51:26 crc kubenswrapper[4967]: E1011 03:51:26.472674 4967 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 11 03:51:26 crc kubenswrapper[4967]: W1011 03:51:26.513281 4967 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout Oct 11 03:51:26 crc kubenswrapper[4967]: I1011 03:51:26.513341 4967 trace.go:236] Trace[1155340602]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (11-Oct-2025 03:51:16.511) (total time: 10001ms): Oct 11 03:51:26 crc kubenswrapper[4967]: Trace[1155340602]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (03:51:26.513) Oct 11 03:51:26 crc kubenswrapper[4967]: Trace[1155340602]: [10.001636399s] [10.001636399s] END Oct 11 03:51:26 crc kubenswrapper[4967]: E1011 03:51:26.513357 4967 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Oct 11 03:51:26 crc kubenswrapper[4967]: I1011 03:51:26.658882 4967 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 11 03:51:26 crc kubenswrapper[4967]: I1011 03:51:26.658966 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 11 03:51:26 crc kubenswrapper[4967]: I1011 03:51:26.730804 4967 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 11 03:51:26 crc kubenswrapper[4967]: I1011 03:51:26.730870 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 11 03:51:26 crc kubenswrapper[4967]: I1011 03:51:26.742966 4967 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Oct 11 03:51:26 crc kubenswrapper[4967]: I1011 03:51:26.743052 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Oct 11 03:51:29 crc kubenswrapper[4967]: I1011 03:51:29.400782 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:51:29 crc kubenswrapper[4967]: I1011 03:51:29.401130 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:29 crc kubenswrapper[4967]: I1011 03:51:29.402937 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:29 crc kubenswrapper[4967]: I1011 03:51:29.402995 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:29 crc kubenswrapper[4967]: I1011 03:51:29.403009 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:29 crc kubenswrapper[4967]: I1011 03:51:29.405980 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:51:29 crc kubenswrapper[4967]: I1011 03:51:29.924723 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:29 crc kubenswrapper[4967]: I1011 03:51:29.925975 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:29 crc kubenswrapper[4967]: I1011 03:51:29.926034 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:29 crc kubenswrapper[4967]: I1011 03:51:29.926052 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.456212 4967 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.742463 4967 apiserver.go:52] "Watching apiserver" Oct 11 03:51:31 crc kubenswrapper[4967]: E1011 03:51:31.742807 4967 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": context deadline exceeded" interval="6.4s" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.745270 4967 trace.go:236] Trace[31587912]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (11-Oct-2025 03:51:20.214) (total time: 11530ms): Oct 11 03:51:31 crc kubenswrapper[4967]: Trace[31587912]: ---"Objects listed" error: 11530ms (03:51:31.745) Oct 11 03:51:31 crc kubenswrapper[4967]: Trace[31587912]: [11.530979771s] [11.530979771s] END Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.745360 4967 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.747404 4967 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Oct 11 03:51:31 crc kubenswrapper[4967]: E1011 03:51:31.747803 4967 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes \"crc\" is forbidden: autoscaling.openshift.io/ManagedNode infra config cache not synchronized" node="crc" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.748426 4967 trace.go:236] Trace[2094923370]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (11-Oct-2025 03:51:16.844) (total time: 14904ms): Oct 11 03:51:31 crc kubenswrapper[4967]: Trace[2094923370]: ---"Objects listed" error: 14904ms (03:51:31.748) Oct 11 03:51:31 crc kubenswrapper[4967]: Trace[2094923370]: [14.904218637s] [14.904218637s] END Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.748459 4967 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.749888 4967 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.750203 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-diagnostics/network-check-target-xd92c","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h","openshift-network-operator/network-operator-58b4c7f79c-55gtf"] Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.750947 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.751662 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.751724 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.751755 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.751781 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:51:31 crc kubenswrapper[4967]: E1011 03:51:31.752217 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:51:31 crc kubenswrapper[4967]: E1011 03:51:31.753326 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.755103 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.756236 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.756237 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.756342 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Oct 11 03:51:31 crc kubenswrapper[4967]: E1011 03:51:31.756414 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.758385 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.758690 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.758762 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.759154 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.759518 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.759670 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.784862 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.799503 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.818207 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.830923 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.848006 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.848101 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.848141 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.848168 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.848217 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.848244 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.848279 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.848317 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.848349 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.848384 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.848415 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.848441 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.848471 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.848557 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:51:31 crc kubenswrapper[4967]: E1011 03:51:31.848671 4967 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 11 03:51:31 crc kubenswrapper[4967]: E1011 03:51:31.848766 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-11 03:51:32.348741714 +0000 UTC m=+20.311950827 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.848870 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 11 03:51:31 crc kubenswrapper[4967]: E1011 03:51:31.849039 4967 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 11 03:51:31 crc kubenswrapper[4967]: E1011 03:51:31.849135 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-11 03:51:32.349115572 +0000 UTC m=+20.312324675 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.849432 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.849542 4967 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.854495 4967 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.858321 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.858546 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.859204 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 11 03:51:31 crc kubenswrapper[4967]: E1011 03:51:31.870209 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 11 03:51:31 crc kubenswrapper[4967]: E1011 03:51:31.870245 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 11 03:51:31 crc kubenswrapper[4967]: E1011 03:51:31.870261 4967 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:51:31 crc kubenswrapper[4967]: E1011 03:51:31.870561 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-11 03:51:32.370345464 +0000 UTC m=+20.333554397 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.875628 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 11 03:51:31 crc kubenswrapper[4967]: E1011 03:51:31.876336 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 11 03:51:31 crc kubenswrapper[4967]: E1011 03:51:31.876385 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 11 03:51:31 crc kubenswrapper[4967]: E1011 03:51:31.876404 4967 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:51:31 crc kubenswrapper[4967]: E1011 03:51:31.876490 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-11 03:51:32.376461027 +0000 UTC m=+20.339669960 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.880022 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.882533 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.882540 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.892954 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.902707 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.928687 4967 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:59562->192.168.126.11:17697: read: connection reset by peer" start-of-body= Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.928776 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": read tcp 192.168.126.11:59562->192.168.126.11:17697: read: connection reset by peer" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.929314 4967 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.929362 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.929658 4967 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver-check-endpoints namespace/openshift-kube-apiserver: Readiness probe status=failure output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" start-of-body= Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.929714 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" probeResult="failure" output="Get \"https://192.168.126.11:17697/healthz\": dial tcp 192.168.126.11:17697: connect: connection refused" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.949382 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.949430 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.949488 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.949517 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.949539 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.949574 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.949593 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.949613 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.949633 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.949654 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.949675 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.949698 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.949736 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.949763 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.949786 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.949810 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.949831 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.949852 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.949873 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.949895 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.949918 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.949938 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.949959 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.949987 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950013 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950034 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950057 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950382 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950406 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950427 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950454 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950475 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950498 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950500 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950520 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950542 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950584 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950606 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950627 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950647 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950667 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950690 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950710 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950716 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950730 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950752 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950774 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950798 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950819 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950840 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950862 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950883 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950903 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950908 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950927 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950948 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950971 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.950992 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951012 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951032 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951054 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951095 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951115 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951138 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951159 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951179 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951201 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951222 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951244 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951298 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951322 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951344 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951367 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951390 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951413 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951435 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951458 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951485 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951506 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951530 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951552 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951575 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951597 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951619 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951641 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951663 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951684 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951707 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951730 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951793 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951818 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951842 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951864 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951887 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951912 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951936 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951959 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951983 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952006 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952028 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952052 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952133 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952158 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952180 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952203 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952224 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952248 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952270 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952293 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952318 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952341 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952364 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952386 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952408 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952429 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952454 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952476 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952497 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952519 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952543 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952565 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952586 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952609 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952633 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952655 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952676 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952699 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952722 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952744 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952769 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952794 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952816 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952838 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952860 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952969 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952996 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953020 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953045 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953088 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953113 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953136 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953160 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953184 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953209 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953232 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953255 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953278 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953302 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953326 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953351 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953375 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953399 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953424 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953446 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953471 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953506 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953531 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953555 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953579 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953603 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953627 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953651 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953675 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953699 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953723 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953746 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953770 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953796 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953821 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953844 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953868 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953891 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953916 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953944 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953970 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953995 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.954018 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.954041 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.954065 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.954108 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.954133 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.954159 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.954185 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.954208 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.954230 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.954254 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.954278 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.954302 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.954325 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.954349 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.954371 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.954399 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.954424 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.954450 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.954475 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.954500 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.954523 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.954545 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.954593 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.954717 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.954764 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.954781 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.954797 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.954844 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.951116 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952157 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952331 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952791 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.952957 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953430 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.953881 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.954359 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.954534 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.954712 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.954905 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.955617 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.955098 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.955111 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.955294 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.955364 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.955911 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.956038 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.956210 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.956271 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.956539 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.956701 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.956838 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.956862 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.956845 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.957012 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.957031 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.957086 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.957103 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.957242 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.957246 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.957322 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.957513 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.957589 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.957654 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.957832 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.957880 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.958573 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.958730 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.958992 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.959091 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.959210 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.959262 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.959351 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.959466 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.959601 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.959640 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.959693 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.959801 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.959914 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.960208 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.960332 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.960393 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.960448 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.960337 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.960780 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.961941 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.962023 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.962028 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.962251 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.962271 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.962308 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.962389 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.962417 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.962545 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.962866 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.962924 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.963373 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.962421 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.963419 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.963459 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.963488 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.963602 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.963700 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.963715 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.963811 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.963876 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.963894 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.964017 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.964181 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.964189 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.964355 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.964466 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.964548 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.964553 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.964883 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.965186 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.965898 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.965974 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.966779 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.967053 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.967154 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.967309 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.967416 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.967583 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.967785 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.968034 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.968249 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.968412 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.968571 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.968666 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.968732 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.969411 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.969499 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.969544 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.969587 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.969594 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.969721 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.969768 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.969852 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.970635 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.970821 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.970969 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.971216 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.971675 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.971793 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.971810 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.971936 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.971358 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.972058 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.972102 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.972171 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.972686 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.972836 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.973327 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.973521 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.973557 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.973678 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.973719 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.974049 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.974243 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.974363 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.974416 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.974494 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.974646 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.975148 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.975205 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.975314 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.975590 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.975606 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.975851 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.975935 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.975944 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.976009 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.976162 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.976212 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.976319 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.976507 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.976521 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.976539 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.976619 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.976701 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.976779 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: E1011 03:51:31.977319 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:51:32.47546856 +0000 UTC m=+20.438677483 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.977381 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.977822 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.977990 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.979769 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.979934 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.979975 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.980122 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.980136 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.980817 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.980931 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.981036 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.982941 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.982959 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.983020 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.983043 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.983067 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.983155 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.983337 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.983935 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.984102 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.984677 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.985106 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.985934 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.986034 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.986186 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.986315 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.986325 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.986412 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.986477 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.986615 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.986712 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.986820 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.990369 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.990781 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:31 crc kubenswrapper[4967]: I1011 03:51:31.993921 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.004685 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.012270 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.018148 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.021482 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056183 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056217 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056229 4967 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056239 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056249 4967 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056259 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056267 4967 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056277 4967 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056285 4967 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056293 4967 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056300 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056308 4967 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056316 4967 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056324 4967 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056332 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056340 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056347 4967 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056355 4967 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056364 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056395 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056404 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056414 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056422 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056430 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056438 4967 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056446 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056455 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056463 4967 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056471 4967 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056479 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056487 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056495 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056505 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056513 4967 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056521 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056529 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056537 4967 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056545 4967 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056554 4967 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056561 4967 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056569 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056578 4967 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056585 4967 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056593 4967 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056602 4967 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056610 4967 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056618 4967 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056627 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056635 4967 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056643 4967 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056651 4967 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056660 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056668 4967 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056675 4967 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056683 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056690 4967 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056699 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056706 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056714 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056723 4967 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056731 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056739 4967 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056747 4967 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056756 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056764 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056772 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056780 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056789 4967 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056798 4967 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056806 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056814 4967 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056823 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056832 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056841 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056849 4967 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056858 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056866 4967 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056873 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056881 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056889 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056897 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056904 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056912 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056920 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056928 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056936 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056945 4967 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.056953 4967 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057118 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057131 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057141 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057151 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057160 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057170 4967 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057179 4967 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057187 4967 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057194 4967 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057202 4967 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057210 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057219 4967 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057226 4967 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057234 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057243 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057250 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057258 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057267 4967 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057275 4967 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057283 4967 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057290 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057298 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057310 4967 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057319 4967 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057328 4967 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057337 4967 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057345 4967 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057352 4967 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057361 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057369 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057377 4967 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057386 4967 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057394 4967 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057401 4967 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057409 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057418 4967 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057425 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057433 4967 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057441 4967 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057449 4967 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057456 4967 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057465 4967 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057473 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057481 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057489 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057497 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057504 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057513 4967 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057521 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057529 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057537 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057545 4967 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057553 4967 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057561 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057571 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057580 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057588 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057597 4967 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057605 4967 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057613 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057621 4967 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057629 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057637 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057645 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057654 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057663 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057671 4967 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057680 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057688 4967 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057696 4967 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057704 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057712 4967 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057720 4967 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057728 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057736 4967 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057745 4967 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057753 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057760 4967 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057769 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057777 4967 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057785 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057793 4967 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057802 4967 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057810 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057819 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057827 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057836 4967 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057844 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057854 4967 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057862 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057870 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057878 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057886 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057894 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057901 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057909 4967 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057917 4967 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057925 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057934 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057941 4967 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057948 4967 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057956 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.057965 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.076204 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.087812 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Oct 11 03:51:32 crc kubenswrapper[4967]: W1011 03:51:32.091497 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-1fd0b1bf7f8a9a497fff6d986b289cbefe6cd89091f8e2187141d162013735c7 WatchSource:0}: Error finding container 1fd0b1bf7f8a9a497fff6d986b289cbefe6cd89091f8e2187141d162013735c7: Status 404 returned error can't find the container with id 1fd0b1bf7f8a9a497fff6d986b289cbefe6cd89091f8e2187141d162013735c7 Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.097149 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Oct 11 03:51:32 crc kubenswrapper[4967]: W1011 03:51:32.102605 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-5a48e40f46b9a01ad1dd1bd84c58bc19b256dd15091291d52ae0cf8c9ecd6bc3 WatchSource:0}: Error finding container 5a48e40f46b9a01ad1dd1bd84c58bc19b256dd15091291d52ae0cf8c9ecd6bc3: Status 404 returned error can't find the container with id 5a48e40f46b9a01ad1dd1bd84c58bc19b256dd15091291d52ae0cf8c9ecd6bc3 Oct 11 03:51:32 crc kubenswrapper[4967]: W1011 03:51:32.111756 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-8f8568bd9b97f8af57a414e5fa0036d3dbf9ea0aeff1ffe7bc30439f980ffc34 WatchSource:0}: Error finding container 8f8568bd9b97f8af57a414e5fa0036d3dbf9ea0aeff1ffe7bc30439f980ffc34: Status 404 returned error can't find the container with id 8f8568bd9b97f8af57a414e5fa0036d3dbf9ea0aeff1ffe7bc30439f980ffc34 Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.360377 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.360534 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:51:32 crc kubenswrapper[4967]: E1011 03:51:32.360603 4967 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 11 03:51:32 crc kubenswrapper[4967]: E1011 03:51:32.360704 4967 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 11 03:51:32 crc kubenswrapper[4967]: E1011 03:51:32.360784 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-11 03:51:33.36075367 +0000 UTC m=+21.323962613 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 11 03:51:32 crc kubenswrapper[4967]: E1011 03:51:32.360821 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-11 03:51:33.360806551 +0000 UTC m=+21.324015494 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.462532 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.462608 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:51:32 crc kubenswrapper[4967]: E1011 03:51:32.462797 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 11 03:51:32 crc kubenswrapper[4967]: E1011 03:51:32.462833 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 11 03:51:32 crc kubenswrapper[4967]: E1011 03:51:32.462846 4967 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:51:32 crc kubenswrapper[4967]: E1011 03:51:32.462932 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-11 03:51:33.462915292 +0000 UTC m=+21.426124225 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:51:32 crc kubenswrapper[4967]: E1011 03:51:32.463011 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 11 03:51:32 crc kubenswrapper[4967]: E1011 03:51:32.463117 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 11 03:51:32 crc kubenswrapper[4967]: E1011 03:51:32.463131 4967 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:51:32 crc kubenswrapper[4967]: E1011 03:51:32.463168 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-11 03:51:33.463158697 +0000 UTC m=+21.426367630 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.562881 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:51:32 crc kubenswrapper[4967]: E1011 03:51:32.563097 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:51:33.563025099 +0000 UTC m=+21.526234032 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.775996 4967 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.817401 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.817891 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.819097 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.819703 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.820633 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.821119 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.821654 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.822539 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.823155 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.824030 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.824489 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.825460 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.825909 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.826409 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.826563 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.827326 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.827807 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.829246 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.830034 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.831023 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.832744 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.833497 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.835304 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.835913 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.837630 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.838318 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.838959 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.839586 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.839535 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.840045 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.840596 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.841028 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.841488 4967 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.841580 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.842928 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.843445 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.843825 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.844884 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.845522 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.846011 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.849402 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.850030 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.850943 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.851619 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.852581 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.853145 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.853178 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.853985 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.854525 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.855392 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.856191 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.856971 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.857438 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.858244 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.858723 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.859390 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.860298 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.869472 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.886625 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.898008 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.936640 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb"} Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.936699 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"5a48e40f46b9a01ad1dd1bd84c58bc19b256dd15091291d52ae0cf8c9ecd6bc3"} Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.938529 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"1fd0b1bf7f8a9a497fff6d986b289cbefe6cd89091f8e2187141d162013735c7"} Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.940024 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.942035 4967 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42" exitCode=255 Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.942104 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42"} Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.944102 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c"} Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.944138 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526"} Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.944147 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"8f8568bd9b97f8af57a414e5fa0036d3dbf9ea0aeff1ffe7bc30439f980ffc34"} Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.953907 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.956609 4967 scope.go:117] "RemoveContainer" containerID="1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.957435 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.971660 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:32 crc kubenswrapper[4967]: I1011 03:51:32.985945 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.019436 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.043477 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.046596 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.064499 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.086284 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.120797 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/node-resolver-jw555"] Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.121147 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-jw555" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.127099 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.127366 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.131365 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.137417 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.156083 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.168163 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/84b4af1b-44af-4ae1-9687-7790849c7d56-hosts-file\") pod \"node-resolver-jw555\" (UID: \"84b4af1b-44af-4ae1-9687-7790849c7d56\") " pod="openshift-dns/node-resolver-jw555" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.168252 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-95dhf\" (UniqueName: \"kubernetes.io/projected/84b4af1b-44af-4ae1-9687-7790849c7d56-kube-api-access-95dhf\") pod \"node-resolver-jw555\" (UID: \"84b4af1b-44af-4ae1-9687-7790849c7d56\") " pod="openshift-dns/node-resolver-jw555" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.177305 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.220455 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.237159 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.253861 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.268458 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.268549 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-95dhf\" (UniqueName: \"kubernetes.io/projected/84b4af1b-44af-4ae1-9687-7790849c7d56-kube-api-access-95dhf\") pod \"node-resolver-jw555\" (UID: \"84b4af1b-44af-4ae1-9687-7790849c7d56\") " pod="openshift-dns/node-resolver-jw555" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.268922 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/84b4af1b-44af-4ae1-9687-7790849c7d56-hosts-file\") pod \"node-resolver-jw555\" (UID: \"84b4af1b-44af-4ae1-9687-7790849c7d56\") " pod="openshift-dns/node-resolver-jw555" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.268981 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/84b4af1b-44af-4ae1-9687-7790849c7d56-hosts-file\") pod \"node-resolver-jw555\" (UID: \"84b4af1b-44af-4ae1-9687-7790849c7d56\") " pod="openshift-dns/node-resolver-jw555" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.282881 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.287012 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-95dhf\" (UniqueName: \"kubernetes.io/projected/84b4af1b-44af-4ae1-9687-7790849c7d56-kube-api-access-95dhf\") pod \"node-resolver-jw555\" (UID: \"84b4af1b-44af-4ae1-9687-7790849c7d56\") " pod="openshift-dns/node-resolver-jw555" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.295185 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.307919 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.320180 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.331576 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.342458 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.361605 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.370090 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.370146 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:51:33 crc kubenswrapper[4967]: E1011 03:51:33.370208 4967 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 11 03:51:33 crc kubenswrapper[4967]: E1011 03:51:33.370263 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-11 03:51:35.370250097 +0000 UTC m=+23.333459030 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 11 03:51:33 crc kubenswrapper[4967]: E1011 03:51:33.370319 4967 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 11 03:51:33 crc kubenswrapper[4967]: E1011 03:51:33.370399 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-11 03:51:35.37038204 +0000 UTC m=+23.333590973 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.380067 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.390739 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.399409 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.440212 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-jw555" Oct 11 03:51:33 crc kubenswrapper[4967]: W1011 03:51:33.451267 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod84b4af1b_44af_4ae1_9687_7790849c7d56.slice/crio-b67ffcc3794f64d176974288dcb06707d7594ddb91552453583cc9c64f5bab2e WatchSource:0}: Error finding container b67ffcc3794f64d176974288dcb06707d7594ddb91552453583cc9c64f5bab2e: Status 404 returned error can't find the container with id b67ffcc3794f64d176974288dcb06707d7594ddb91552453583cc9c64f5bab2e Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.470997 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.471035 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:51:33 crc kubenswrapper[4967]: E1011 03:51:33.471160 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 11 03:51:33 crc kubenswrapper[4967]: E1011 03:51:33.471175 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 11 03:51:33 crc kubenswrapper[4967]: E1011 03:51:33.471188 4967 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:51:33 crc kubenswrapper[4967]: E1011 03:51:33.471232 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-11 03:51:35.471219423 +0000 UTC m=+23.434428356 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:51:33 crc kubenswrapper[4967]: E1011 03:51:33.471504 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 11 03:51:33 crc kubenswrapper[4967]: E1011 03:51:33.471515 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 11 03:51:33 crc kubenswrapper[4967]: E1011 03:51:33.471522 4967 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:51:33 crc kubenswrapper[4967]: E1011 03:51:33.471543 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-11 03:51:35.47153689 +0000 UTC m=+23.434745823 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.571723 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:51:33 crc kubenswrapper[4967]: E1011 03:51:33.571906 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:51:35.571881452 +0000 UTC m=+23.535090385 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.665931 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.672626 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.680156 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.694595 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.706275 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.719322 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.734959 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.753649 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.765052 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.786858 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.801846 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.814006 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.814021 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.814086 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:51:33 crc kubenswrapper[4967]: E1011 03:51:33.814188 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:51:33 crc kubenswrapper[4967]: E1011 03:51:33.814330 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:51:33 crc kubenswrapper[4967]: E1011 03:51:33.814435 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.855750 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.884996 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.898824 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-6xvgx"] Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.899204 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-daemon-sgrv2"] Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.899373 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-6xvgx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.899533 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.899771 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-additional-cni-plugins-zwjmt"] Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.902843 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.902911 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.903352 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.903436 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.904270 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.904391 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.904571 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.909424 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.911681 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-lw6dx"] Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.912458 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.915652 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.917137 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.917625 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.917713 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.919765 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.920004 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.920080 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.920688 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.921860 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.921987 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.923985 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.924098 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.924022 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.945028 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.947800 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-jw555" event={"ID":"84b4af1b-44af-4ae1-9687-7790849c7d56","Type":"ContainerStarted","Data":"42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625"} Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.947992 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-jw555" event={"ID":"84b4af1b-44af-4ae1-9687-7790849c7d56","Type":"ContainerStarted","Data":"b67ffcc3794f64d176974288dcb06707d7594ddb91552453583cc9c64f5bab2e"} Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.949431 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.950669 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e"} Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.969799 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.975302 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-run-openvswitch\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.975331 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vqqw5\" (UniqueName: \"kubernetes.io/projected/1db525c3-0479-4585-b107-a448f2b94ffd-kube-api-access-vqqw5\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.975351 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-cni-bin\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.975367 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/270ac8da-b130-4f4f-bbf7-655229b64446-ovnkube-script-lib\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.975382 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8bsk\" (UniqueName: \"kubernetes.io/projected/270ac8da-b130-4f4f-bbf7-655229b64446-kube-api-access-v8bsk\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.975399 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-multus-cni-dir\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.975483 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-run-ovn\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.975524 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-log-socket\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.975550 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-slash\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.975586 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-os-release\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.975614 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-host-var-lib-cni-bin\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.975634 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-etc-openvswitch\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.975651 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-cnibin\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.975666 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-host-run-k8s-cni-cncf-io\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.975691 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-host-var-lib-cni-multus\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.975782 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-run-netns\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.975841 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-run-systemd\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.975857 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-node-log\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.975888 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-host-run-multus-certs\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.975906 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f89f20ee-089e-446a-85c4-933aacdb1eba-os-release\") pod \"multus-additional-cni-plugins-zwjmt\" (UID: \"f89f20ee-089e-446a-85c4-933aacdb1eba\") " pod="openshift-multus/multus-additional-cni-plugins-zwjmt" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.975923 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f89f20ee-089e-446a-85c4-933aacdb1eba-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zwjmt\" (UID: \"f89f20ee-089e-446a-85c4-933aacdb1eba\") " pod="openshift-multus/multus-additional-cni-plugins-zwjmt" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.975939 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1db525c3-0479-4585-b107-a448f2b94ffd-cni-binary-copy\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.975959 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/97c07678-14be-410c-b61f-498cb49bc960-proxy-tls\") pod \"machine-config-daemon-sgrv2\" (UID: \"97c07678-14be-410c-b61f-498cb49bc960\") " pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.975983 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-multus-conf-dir\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.976037 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/f89f20ee-089e-446a-85c4-933aacdb1eba-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zwjmt\" (UID: \"f89f20ee-089e-446a-85c4-933aacdb1eba\") " pod="openshift-multus/multus-additional-cni-plugins-zwjmt" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.976062 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-systemd-units\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.976114 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/270ac8da-b130-4f4f-bbf7-655229b64446-env-overrides\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.976143 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/270ac8da-b130-4f4f-bbf7-655229b64446-ovn-node-metrics-cert\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.976176 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f89f20ee-089e-446a-85c4-933aacdb1eba-cni-binary-copy\") pod \"multus-additional-cni-plugins-zwjmt\" (UID: \"f89f20ee-089e-446a-85c4-933aacdb1eba\") " pod="openshift-multus/multus-additional-cni-plugins-zwjmt" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.976195 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-kubelet\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.976210 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-run-ovn-kubernetes\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.976255 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t5zwk\" (UniqueName: \"kubernetes.io/projected/f89f20ee-089e-446a-85c4-933aacdb1eba-kube-api-access-t5zwk\") pod \"multus-additional-cni-plugins-zwjmt\" (UID: \"f89f20ee-089e-446a-85c4-933aacdb1eba\") " pod="openshift-multus/multus-additional-cni-plugins-zwjmt" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.976276 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-hostroot\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.976295 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/1db525c3-0479-4585-b107-a448f2b94ffd-multus-daemon-config\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.976326 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-var-lib-openvswitch\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.976340 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/97c07678-14be-410c-b61f-498cb49bc960-mcd-auth-proxy-config\") pod \"machine-config-daemon-sgrv2\" (UID: \"97c07678-14be-410c-b61f-498cb49bc960\") " pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.976355 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fr4zj\" (UniqueName: \"kubernetes.io/projected/97c07678-14be-410c-b61f-498cb49bc960-kube-api-access-fr4zj\") pod \"machine-config-daemon-sgrv2\" (UID: \"97c07678-14be-410c-b61f-498cb49bc960\") " pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.976369 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/270ac8da-b130-4f4f-bbf7-655229b64446-ovnkube-config\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.976382 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-host-run-netns\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.976438 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f89f20ee-089e-446a-85c4-933aacdb1eba-system-cni-dir\") pod \"multus-additional-cni-plugins-zwjmt\" (UID: \"f89f20ee-089e-446a-85c4-933aacdb1eba\") " pod="openshift-multus/multus-additional-cni-plugins-zwjmt" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.976457 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-cni-netd\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.976473 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.976490 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f89f20ee-089e-446a-85c4-933aacdb1eba-cnibin\") pod \"multus-additional-cni-plugins-zwjmt\" (UID: \"f89f20ee-089e-446a-85c4-933aacdb1eba\") " pod="openshift-multus/multus-additional-cni-plugins-zwjmt" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.976503 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/97c07678-14be-410c-b61f-498cb49bc960-rootfs\") pod \"machine-config-daemon-sgrv2\" (UID: \"97c07678-14be-410c-b61f-498cb49bc960\") " pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.976556 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-system-cni-dir\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.976617 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-host-var-lib-kubelet\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.976638 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-etc-kubernetes\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:33 crc kubenswrapper[4967]: I1011 03:51:33.976664 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-multus-socket-dir-parent\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.010086 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:34Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.021601 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:34Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.034136 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:34Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.047027 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:34Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.061472 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:34Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.073706 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:34Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.077682 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-run-ovn\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.077755 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-log-socket\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.077766 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-run-ovn\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.077788 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-slash\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.077833 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-os-release\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.077857 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-slash\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.077871 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-host-var-lib-cni-bin\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.077882 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-log-socket\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.077905 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-etc-openvswitch\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.077975 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-etc-openvswitch\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.078000 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-run-netns\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.077981 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-run-netns\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.078008 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-host-var-lib-cni-bin\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.078041 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-run-systemd\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.078106 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-node-log\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.078106 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-run-systemd\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.078148 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-cnibin\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.078164 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-node-log\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.078179 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-host-run-k8s-cni-cncf-io\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.078206 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-cnibin\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.078215 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-host-var-lib-cni-multus\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.078184 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-os-release\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.078253 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-host-run-multus-certs\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.078302 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-host-run-multus-certs\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.078310 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-host-run-k8s-cni-cncf-io\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.078308 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f89f20ee-089e-446a-85c4-933aacdb1eba-os-release\") pod \"multus-additional-cni-plugins-zwjmt\" (UID: \"f89f20ee-089e-446a-85c4-933aacdb1eba\") " pod="openshift-multus/multus-additional-cni-plugins-zwjmt" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.078348 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f89f20ee-089e-446a-85c4-933aacdb1eba-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zwjmt\" (UID: \"f89f20ee-089e-446a-85c4-933aacdb1eba\") " pod="openshift-multus/multus-additional-cni-plugins-zwjmt" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.078309 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-host-var-lib-cni-multus\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.078374 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1db525c3-0479-4585-b107-a448f2b94ffd-cni-binary-copy\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.078368 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/f89f20ee-089e-446a-85c4-933aacdb1eba-os-release\") pod \"multus-additional-cni-plugins-zwjmt\" (UID: \"f89f20ee-089e-446a-85c4-933aacdb1eba\") " pod="openshift-multus/multus-additional-cni-plugins-zwjmt" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.078412 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/f89f20ee-089e-446a-85c4-933aacdb1eba-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zwjmt\" (UID: \"f89f20ee-089e-446a-85c4-933aacdb1eba\") " pod="openshift-multus/multus-additional-cni-plugins-zwjmt" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.078427 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-systemd-units\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.078451 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-systemd-units\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.078459 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/270ac8da-b130-4f4f-bbf7-655229b64446-env-overrides\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.078501 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/270ac8da-b130-4f4f-bbf7-655229b64446-ovn-node-metrics-cert\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.078535 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/97c07678-14be-410c-b61f-498cb49bc960-proxy-tls\") pod \"machine-config-daemon-sgrv2\" (UID: \"97c07678-14be-410c-b61f-498cb49bc960\") " pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.078564 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-multus-conf-dir\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.078601 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f89f20ee-089e-446a-85c4-933aacdb1eba-cni-binary-copy\") pod \"multus-additional-cni-plugins-zwjmt\" (UID: \"f89f20ee-089e-446a-85c4-933aacdb1eba\") " pod="openshift-multus/multus-additional-cni-plugins-zwjmt" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.078739 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-multus-conf-dir\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.078791 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-kubelet\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.079046 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/270ac8da-b130-4f4f-bbf7-655229b64446-env-overrides\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.079085 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/f89f20ee-089e-446a-85c4-933aacdb1eba-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-zwjmt\" (UID: \"f89f20ee-089e-446a-85c4-933aacdb1eba\") " pod="openshift-multus/multus-additional-cni-plugins-zwjmt" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.079110 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/f89f20ee-089e-446a-85c4-933aacdb1eba-tuning-conf-dir\") pod \"multus-additional-cni-plugins-zwjmt\" (UID: \"f89f20ee-089e-446a-85c4-933aacdb1eba\") " pod="openshift-multus/multus-additional-cni-plugins-zwjmt" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.079323 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/1db525c3-0479-4585-b107-a448f2b94ffd-cni-binary-copy\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.079629 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/f89f20ee-089e-446a-85c4-933aacdb1eba-cni-binary-copy\") pod \"multus-additional-cni-plugins-zwjmt\" (UID: \"f89f20ee-089e-446a-85c4-933aacdb1eba\") " pod="openshift-multus/multus-additional-cni-plugins-zwjmt" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.079727 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-kubelet\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.079777 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-run-ovn-kubernetes\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.079835 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t5zwk\" (UniqueName: \"kubernetes.io/projected/f89f20ee-089e-446a-85c4-933aacdb1eba-kube-api-access-t5zwk\") pod \"multus-additional-cni-plugins-zwjmt\" (UID: \"f89f20ee-089e-446a-85c4-933aacdb1eba\") " pod="openshift-multus/multus-additional-cni-plugins-zwjmt" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.079874 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-hostroot\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.079908 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/1db525c3-0479-4585-b107-a448f2b94ffd-multus-daemon-config\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.079954 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-var-lib-openvswitch\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.079992 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/97c07678-14be-410c-b61f-498cb49bc960-mcd-auth-proxy-config\") pod \"machine-config-daemon-sgrv2\" (UID: \"97c07678-14be-410c-b61f-498cb49bc960\") " pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.080028 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fr4zj\" (UniqueName: \"kubernetes.io/projected/97c07678-14be-410c-b61f-498cb49bc960-kube-api-access-fr4zj\") pod \"machine-config-daemon-sgrv2\" (UID: \"97c07678-14be-410c-b61f-498cb49bc960\") " pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.080095 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f89f20ee-089e-446a-85c4-933aacdb1eba-system-cni-dir\") pod \"multus-additional-cni-plugins-zwjmt\" (UID: \"f89f20ee-089e-446a-85c4-933aacdb1eba\") " pod="openshift-multus/multus-additional-cni-plugins-zwjmt" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.080250 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-run-ovn-kubernetes\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.080262 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-hostroot\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.080299 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-var-lib-openvswitch\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.080421 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/f89f20ee-089e-446a-85c4-933aacdb1eba-system-cni-dir\") pod \"multus-additional-cni-plugins-zwjmt\" (UID: \"f89f20ee-089e-446a-85c4-933aacdb1eba\") " pod="openshift-multus/multus-additional-cni-plugins-zwjmt" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.080134 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-cni-netd\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.080669 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-cni-netd\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.080739 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.080760 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/1db525c3-0479-4585-b107-a448f2b94ffd-multus-daemon-config\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.080786 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/270ac8da-b130-4f4f-bbf7-655229b64446-ovnkube-config\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.080823 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-host-run-netns\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.080842 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.080871 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f89f20ee-089e-446a-85c4-933aacdb1eba-cnibin\") pod \"multus-additional-cni-plugins-zwjmt\" (UID: \"f89f20ee-089e-446a-85c4-933aacdb1eba\") " pod="openshift-multus/multus-additional-cni-plugins-zwjmt" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.080907 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/97c07678-14be-410c-b61f-498cb49bc960-rootfs\") pod \"machine-config-daemon-sgrv2\" (UID: \"97c07678-14be-410c-b61f-498cb49bc960\") " pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.080939 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-system-cni-dir\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.080951 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/97c07678-14be-410c-b61f-498cb49bc960-mcd-auth-proxy-config\") pod \"machine-config-daemon-sgrv2\" (UID: \"97c07678-14be-410c-b61f-498cb49bc960\") " pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.080971 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-host-var-lib-kubelet\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.081005 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/f89f20ee-089e-446a-85c4-933aacdb1eba-cnibin\") pod \"multus-additional-cni-plugins-zwjmt\" (UID: \"f89f20ee-089e-446a-85c4-933aacdb1eba\") " pod="openshift-multus/multus-additional-cni-plugins-zwjmt" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.081005 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-etc-kubernetes\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.081050 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-multus-socket-dir-parent\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.081103 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-etc-kubernetes\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.081146 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-host-run-netns\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.081149 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-run-openvswitch\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.081209 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-multus-socket-dir-parent\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.081264 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-system-cni-dir\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.081260 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-host-var-lib-kubelet\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.081112 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-run-openvswitch\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.081301 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/97c07678-14be-410c-b61f-498cb49bc960-rootfs\") pod \"machine-config-daemon-sgrv2\" (UID: \"97c07678-14be-410c-b61f-498cb49bc960\") " pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.081310 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-cni-bin\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.081336 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-cni-bin\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.081350 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/270ac8da-b130-4f4f-bbf7-655229b64446-ovnkube-script-lib\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.081409 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8bsk\" (UniqueName: \"kubernetes.io/projected/270ac8da-b130-4f4f-bbf7-655229b64446-kube-api-access-v8bsk\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.081430 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/270ac8da-b130-4f4f-bbf7-655229b64446-ovnkube-config\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.081438 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-multus-cni-dir\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.081477 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vqqw5\" (UniqueName: \"kubernetes.io/projected/1db525c3-0479-4585-b107-a448f2b94ffd-kube-api-access-vqqw5\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.081624 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/1db525c3-0479-4585-b107-a448f2b94ffd-multus-cni-dir\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.082001 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/270ac8da-b130-4f4f-bbf7-655229b64446-ovnkube-script-lib\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.083024 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/97c07678-14be-410c-b61f-498cb49bc960-proxy-tls\") pod \"machine-config-daemon-sgrv2\" (UID: \"97c07678-14be-410c-b61f-498cb49bc960\") " pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.086668 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/270ac8da-b130-4f4f-bbf7-655229b64446-ovn-node-metrics-cert\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.096023 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:34Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.100720 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8bsk\" (UniqueName: \"kubernetes.io/projected/270ac8da-b130-4f4f-bbf7-655229b64446-kube-api-access-v8bsk\") pod \"ovnkube-node-lw6dx\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.101777 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t5zwk\" (UniqueName: \"kubernetes.io/projected/f89f20ee-089e-446a-85c4-933aacdb1eba-kube-api-access-t5zwk\") pod \"multus-additional-cni-plugins-zwjmt\" (UID: \"f89f20ee-089e-446a-85c4-933aacdb1eba\") " pod="openshift-multus/multus-additional-cni-plugins-zwjmt" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.103974 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vqqw5\" (UniqueName: \"kubernetes.io/projected/1db525c3-0479-4585-b107-a448f2b94ffd-kube-api-access-vqqw5\") pod \"multus-6xvgx\" (UID: \"1db525c3-0479-4585-b107-a448f2b94ffd\") " pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.110315 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:34Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.113275 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fr4zj\" (UniqueName: \"kubernetes.io/projected/97c07678-14be-410c-b61f-498cb49bc960-kube-api-access-fr4zj\") pod \"machine-config-daemon-sgrv2\" (UID: \"97c07678-14be-410c-b61f-498cb49bc960\") " pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.124566 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:34Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.135240 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:34Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.146441 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:34Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.159786 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:34Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.178848 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:34Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.191029 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:34Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.207045 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:34Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.214875 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-6xvgx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.220995 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.222268 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:34Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.229140 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.235497 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.235514 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:34Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:34 crc kubenswrapper[4967]: W1011 03:51:34.251989 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf89f20ee_089e_446a_85c4_933aacdb1eba.slice/crio-e6d6f2fc9f01b413a149dd66d1153fd126dd208df2e23c1beb504619ee8084fd WatchSource:0}: Error finding container e6d6f2fc9f01b413a149dd66d1153fd126dd208df2e23c1beb504619ee8084fd: Status 404 returned error can't find the container with id e6d6f2fc9f01b413a149dd66d1153fd126dd208df2e23c1beb504619ee8084fd Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.255360 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:34Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:34 crc kubenswrapper[4967]: W1011 03:51:34.259210 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod270ac8da_b130_4f4f_bbf7_655229b64446.slice/crio-71dca86d3ee54432a3df74b4cc1c3b83b07d1fcc602f810a6211f748cf7b3833 WatchSource:0}: Error finding container 71dca86d3ee54432a3df74b4cc1c3b83b07d1fcc602f810a6211f748cf7b3833: Status 404 returned error can't find the container with id 71dca86d3ee54432a3df74b4cc1c3b83b07d1fcc602f810a6211f748cf7b3833 Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.271335 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:34Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.290746 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:34Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.957011 4967 generic.go:334] "Generic (PLEG): container finished" podID="270ac8da-b130-4f4f-bbf7-655229b64446" containerID="d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1" exitCode=0 Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.957118 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" event={"ID":"270ac8da-b130-4f4f-bbf7-655229b64446","Type":"ContainerDied","Data":"d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1"} Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.957171 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" event={"ID":"270ac8da-b130-4f4f-bbf7-655229b64446","Type":"ContainerStarted","Data":"71dca86d3ee54432a3df74b4cc1c3b83b07d1fcc602f810a6211f748cf7b3833"} Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.959756 4967 generic.go:334] "Generic (PLEG): container finished" podID="f89f20ee-089e-446a-85c4-933aacdb1eba" containerID="eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff" exitCode=0 Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.959839 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" event={"ID":"f89f20ee-089e-446a-85c4-933aacdb1eba","Type":"ContainerDied","Data":"eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff"} Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.959896 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" event={"ID":"f89f20ee-089e-446a-85c4-933aacdb1eba","Type":"ContainerStarted","Data":"e6d6f2fc9f01b413a149dd66d1153fd126dd208df2e23c1beb504619ee8084fd"} Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.962069 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" event={"ID":"97c07678-14be-410c-b61f-498cb49bc960","Type":"ContainerStarted","Data":"642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555"} Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.962136 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" event={"ID":"97c07678-14be-410c-b61f-498cb49bc960","Type":"ContainerStarted","Data":"a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb"} Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.962158 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" event={"ID":"97c07678-14be-410c-b61f-498cb49bc960","Type":"ContainerStarted","Data":"c4c89bf87bbfc19af43823fa111dc15f77ab69462948d6b51ef68f37873a9553"} Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.963761 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6xvgx" event={"ID":"1db525c3-0479-4585-b107-a448f2b94ffd","Type":"ContainerStarted","Data":"a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0"} Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.963807 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6xvgx" event={"ID":"1db525c3-0479-4585-b107-a448f2b94ffd","Type":"ContainerStarted","Data":"49be6a792a3ee53da1b471a564466874f6bfbb5891e6dc20874510b74406b208"} Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.965356 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489"} Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.965740 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.980898 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:34Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:34 crc kubenswrapper[4967]: I1011 03:51:34.999054 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:34Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.016626 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:35Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.036947 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:35Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.050570 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:35Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.075758 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:35Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.088095 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:35Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.101512 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:35Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.118638 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:35Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.133453 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:35Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.151830 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:35Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.164136 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:35Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.177725 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:35Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.190983 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:35Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.208470 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:35Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.222463 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:35Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.238456 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:35Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.253821 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:35Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.273676 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:35Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.286304 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:35Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.301445 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:35Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.316912 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:35Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.334339 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:35Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.346621 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:35Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.369321 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:35Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.391421 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:35Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.393790 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.393849 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:51:35 crc kubenswrapper[4967]: E1011 03:51:35.393999 4967 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 11 03:51:35 crc kubenswrapper[4967]: E1011 03:51:35.394125 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-11 03:51:39.394055725 +0000 UTC m=+27.357264668 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 11 03:51:35 crc kubenswrapper[4967]: E1011 03:51:35.393996 4967 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 11 03:51:35 crc kubenswrapper[4967]: E1011 03:51:35.394187 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-11 03:51:39.394175957 +0000 UTC m=+27.357384900 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.409223 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:35Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.421537 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:35Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.494237 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.494294 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:51:35 crc kubenswrapper[4967]: E1011 03:51:35.494392 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 11 03:51:35 crc kubenswrapper[4967]: E1011 03:51:35.494417 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 11 03:51:35 crc kubenswrapper[4967]: E1011 03:51:35.494430 4967 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:51:35 crc kubenswrapper[4967]: E1011 03:51:35.494470 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-11 03:51:39.494456579 +0000 UTC m=+27.457665512 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:51:35 crc kubenswrapper[4967]: E1011 03:51:35.494392 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 11 03:51:35 crc kubenswrapper[4967]: E1011 03:51:35.494506 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 11 03:51:35 crc kubenswrapper[4967]: E1011 03:51:35.494518 4967 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:51:35 crc kubenswrapper[4967]: E1011 03:51:35.494555 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-11 03:51:39.494544811 +0000 UTC m=+27.457753744 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.595318 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:51:35 crc kubenswrapper[4967]: E1011 03:51:35.595518 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:51:39.595491066 +0000 UTC m=+27.558700009 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.814016 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:51:35 crc kubenswrapper[4967]: E1011 03:51:35.814139 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.814149 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.814278 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:51:35 crc kubenswrapper[4967]: E1011 03:51:35.814464 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:51:35 crc kubenswrapper[4967]: E1011 03:51:35.814555 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.972709 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" event={"ID":"270ac8da-b130-4f4f-bbf7-655229b64446","Type":"ContainerStarted","Data":"e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec"} Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.973415 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" event={"ID":"270ac8da-b130-4f4f-bbf7-655229b64446","Type":"ContainerStarted","Data":"faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2"} Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.973491 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" event={"ID":"270ac8da-b130-4f4f-bbf7-655229b64446","Type":"ContainerStarted","Data":"13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a"} Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.973516 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" event={"ID":"270ac8da-b130-4f4f-bbf7-655229b64446","Type":"ContainerStarted","Data":"37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312"} Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.973534 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" event={"ID":"270ac8da-b130-4f4f-bbf7-655229b64446","Type":"ContainerStarted","Data":"963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6"} Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.974774 4967 generic.go:334] "Generic (PLEG): container finished" podID="f89f20ee-089e-446a-85c4-933aacdb1eba" containerID="c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544" exitCode=0 Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.974937 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" event={"ID":"f89f20ee-089e-446a-85c4-933aacdb1eba","Type":"ContainerDied","Data":"c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544"} Oct 11 03:51:35 crc kubenswrapper[4967]: I1011 03:51:35.995451 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:35Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.008598 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:36Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.035431 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:36Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.066229 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:36Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.095928 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:36Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.108021 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:36Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.121684 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:36Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.133328 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:36Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.144327 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:36Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.155752 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:36Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.166508 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:36Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.184702 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:36Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.194819 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:36Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.207651 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:36Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.503391 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/node-ca-48c4c"] Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.503802 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-48c4c" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.506422 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.506439 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.507130 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.508221 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.527951 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:36Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.547986 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:36Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.560602 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:36Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.575553 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:36Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.595789 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:36Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.605917 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5187b2c2-8215-4c7c-8906-8f3d275415a0-serviceca\") pod \"node-ca-48c4c\" (UID: \"5187b2c2-8215-4c7c-8906-8f3d275415a0\") " pod="openshift-image-registry/node-ca-48c4c" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.606042 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7bw8\" (UniqueName: \"kubernetes.io/projected/5187b2c2-8215-4c7c-8906-8f3d275415a0-kube-api-access-m7bw8\") pod \"node-ca-48c4c\" (UID: \"5187b2c2-8215-4c7c-8906-8f3d275415a0\") " pod="openshift-image-registry/node-ca-48c4c" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.606142 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5187b2c2-8215-4c7c-8906-8f3d275415a0-host\") pod \"node-ca-48c4c\" (UID: \"5187b2c2-8215-4c7c-8906-8f3d275415a0\") " pod="openshift-image-registry/node-ca-48c4c" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.610627 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:36Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.624911 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:36Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.642716 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:36Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.658353 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:36Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.672764 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:36Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.686420 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:36Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.705010 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:36Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.707353 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5187b2c2-8215-4c7c-8906-8f3d275415a0-serviceca\") pod \"node-ca-48c4c\" (UID: \"5187b2c2-8215-4c7c-8906-8f3d275415a0\") " pod="openshift-image-registry/node-ca-48c4c" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.707423 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7bw8\" (UniqueName: \"kubernetes.io/projected/5187b2c2-8215-4c7c-8906-8f3d275415a0-kube-api-access-m7bw8\") pod \"node-ca-48c4c\" (UID: \"5187b2c2-8215-4c7c-8906-8f3d275415a0\") " pod="openshift-image-registry/node-ca-48c4c" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.707476 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5187b2c2-8215-4c7c-8906-8f3d275415a0-host\") pod \"node-ca-48c4c\" (UID: \"5187b2c2-8215-4c7c-8906-8f3d275415a0\") " pod="openshift-image-registry/node-ca-48c4c" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.707537 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5187b2c2-8215-4c7c-8906-8f3d275415a0-host\") pod \"node-ca-48c4c\" (UID: \"5187b2c2-8215-4c7c-8906-8f3d275415a0\") " pod="openshift-image-registry/node-ca-48c4c" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.709423 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/5187b2c2-8215-4c7c-8906-8f3d275415a0-serviceca\") pod \"node-ca-48c4c\" (UID: \"5187b2c2-8215-4c7c-8906-8f3d275415a0\") " pod="openshift-image-registry/node-ca-48c4c" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.724487 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:36Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.738297 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:36Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.738725 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7bw8\" (UniqueName: \"kubernetes.io/projected/5187b2c2-8215-4c7c-8906-8f3d275415a0-kube-api-access-m7bw8\") pod \"node-ca-48c4c\" (UID: \"5187b2c2-8215-4c7c-8906-8f3d275415a0\") " pod="openshift-image-registry/node-ca-48c4c" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.755062 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:36Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.828645 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-48c4c" Oct 11 03:51:36 crc kubenswrapper[4967]: W1011 03:51:36.844949 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5187b2c2_8215_4c7c_8906_8f3d275415a0.slice/crio-8f17ddecc281b8120ac6eb1473e1f20e57cfeb8687fe6890e581d231b1b53072 WatchSource:0}: Error finding container 8f17ddecc281b8120ac6eb1473e1f20e57cfeb8687fe6890e581d231b1b53072: Status 404 returned error can't find the container with id 8f17ddecc281b8120ac6eb1473e1f20e57cfeb8687fe6890e581d231b1b53072 Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.981063 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" event={"ID":"270ac8da-b130-4f4f-bbf7-655229b64446","Type":"ContainerStarted","Data":"20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe"} Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.983625 4967 generic.go:334] "Generic (PLEG): container finished" podID="f89f20ee-089e-446a-85c4-933aacdb1eba" containerID="7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a" exitCode=0 Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.983689 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" event={"ID":"f89f20ee-089e-446a-85c4-933aacdb1eba","Type":"ContainerDied","Data":"7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a"} Oct 11 03:51:36 crc kubenswrapper[4967]: I1011 03:51:36.984870 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-48c4c" event={"ID":"5187b2c2-8215-4c7c-8906-8f3d275415a0","Type":"ContainerStarted","Data":"8f17ddecc281b8120ac6eb1473e1f20e57cfeb8687fe6890e581d231b1b53072"} Oct 11 03:51:37 crc kubenswrapper[4967]: I1011 03:51:37.003983 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:37Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:37 crc kubenswrapper[4967]: I1011 03:51:37.017352 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:37Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:37 crc kubenswrapper[4967]: I1011 03:51:37.042357 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:37Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:37 crc kubenswrapper[4967]: I1011 03:51:37.054682 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:37Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:37 crc kubenswrapper[4967]: I1011 03:51:37.066865 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:37Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:37 crc kubenswrapper[4967]: I1011 03:51:37.078902 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:37Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:37 crc kubenswrapper[4967]: I1011 03:51:37.092560 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:37Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:37 crc kubenswrapper[4967]: I1011 03:51:37.104291 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:37Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:37 crc kubenswrapper[4967]: I1011 03:51:37.116213 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:37Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:37 crc kubenswrapper[4967]: I1011 03:51:37.132702 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:37Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:37 crc kubenswrapper[4967]: I1011 03:51:37.146204 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:37Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:37 crc kubenswrapper[4967]: I1011 03:51:37.159371 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:37Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:37 crc kubenswrapper[4967]: I1011 03:51:37.172386 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:37Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:37 crc kubenswrapper[4967]: I1011 03:51:37.184541 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:37Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:37 crc kubenswrapper[4967]: I1011 03:51:37.195494 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:37Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:37 crc kubenswrapper[4967]: I1011 03:51:37.814792 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:51:37 crc kubenswrapper[4967]: I1011 03:51:37.814848 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:51:37 crc kubenswrapper[4967]: I1011 03:51:37.814814 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:51:37 crc kubenswrapper[4967]: E1011 03:51:37.814982 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:51:37 crc kubenswrapper[4967]: E1011 03:51:37.815108 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:51:37 crc kubenswrapper[4967]: E1011 03:51:37.815250 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:51:37 crc kubenswrapper[4967]: I1011 03:51:37.998264 4967 generic.go:334] "Generic (PLEG): container finished" podID="f89f20ee-089e-446a-85c4-933aacdb1eba" containerID="4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036" exitCode=0 Oct 11 03:51:37 crc kubenswrapper[4967]: I1011 03:51:37.998369 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" event={"ID":"f89f20ee-089e-446a-85c4-933aacdb1eba","Type":"ContainerDied","Data":"4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036"} Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.001563 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-48c4c" event={"ID":"5187b2c2-8215-4c7c-8906-8f3d275415a0","Type":"ContainerStarted","Data":"ad406c24472c9ef7da82485a0600cf54f9002b83c0770c2ac86eca0f8779a870"} Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.034986 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.058616 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.075599 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.090835 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.103410 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.112601 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.125696 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.144805 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.148045 4967 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.150315 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.150362 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.150376 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.150511 4967 kubelet_node_status.go:76] "Attempting to register node" node="crc" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.159058 4967 kubelet_node_status.go:115] "Node was previously registered" node="crc" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.159317 4967 kubelet_node_status.go:79] "Successfully registered node" node="crc" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.160292 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.160325 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.160336 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.160352 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.160364 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:38Z","lastTransitionTime":"2025-10-11T03:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.161239 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: E1011 03:51:38.175114 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.176151 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.179239 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.179272 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.179280 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.179294 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.179303 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:38Z","lastTransitionTime":"2025-10-11T03:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.186401 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: E1011 03:51:38.191372 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.195256 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.195319 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.195342 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.195366 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.195386 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:38Z","lastTransitionTime":"2025-10-11T03:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.198768 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: E1011 03:51:38.207963 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.211717 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.212254 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.212311 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.212329 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.212354 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.212374 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:38Z","lastTransitionTime":"2025-10-11T03:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:38 crc kubenswrapper[4967]: E1011 03:51:38.227517 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.229265 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.234766 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.234869 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.234926 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.235010 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.235086 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:38Z","lastTransitionTime":"2025-10-11T03:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.245146 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: E1011 03:51:38.247268 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: E1011 03:51:38.247570 4967 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.249361 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.249388 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.249397 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.249409 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.249418 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:38Z","lastTransitionTime":"2025-10-11T03:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.262167 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.275871 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.288822 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.300102 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.318532 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.331656 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.343062 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.351885 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.351916 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.351929 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.351946 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.351957 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:38Z","lastTransitionTime":"2025-10-11T03:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.354786 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.373505 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.388145 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.406449 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.423576 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.440740 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.454212 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.454249 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.454259 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.454273 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.454283 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:38Z","lastTransitionTime":"2025-10-11T03:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.458957 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.472439 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad406c24472c9ef7da82485a0600cf54f9002b83c0770c2ac86eca0f8779a870\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:38Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.557445 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.557516 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.557533 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.557557 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.557574 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:38Z","lastTransitionTime":"2025-10-11T03:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.660712 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.660770 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.660789 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.660811 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.660829 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:38Z","lastTransitionTime":"2025-10-11T03:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.763672 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.763734 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.763752 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.763776 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.763794 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:38Z","lastTransitionTime":"2025-10-11T03:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.866492 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.866530 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.866539 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.866556 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.866565 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:38Z","lastTransitionTime":"2025-10-11T03:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.969593 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.969642 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.969650 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.969666 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:38 crc kubenswrapper[4967]: I1011 03:51:38.969675 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:38Z","lastTransitionTime":"2025-10-11T03:51:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.009435 4967 generic.go:334] "Generic (PLEG): container finished" podID="f89f20ee-089e-446a-85c4-933aacdb1eba" containerID="723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573" exitCode=0 Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.009507 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" event={"ID":"f89f20ee-089e-446a-85c4-933aacdb1eba","Type":"ContainerDied","Data":"723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573"} Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.016761 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" event={"ID":"270ac8da-b130-4f4f-bbf7-655229b64446","Type":"ContainerStarted","Data":"12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4"} Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.030802 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:39Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.053184 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:39Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.071185 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:39Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.072555 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.072622 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.072641 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.072661 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.072675 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:39Z","lastTransitionTime":"2025-10-11T03:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.086715 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:39Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.099870 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:39Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.124726 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:39Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.138434 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:39Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.151195 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:39Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.164530 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:39Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.175254 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad406c24472c9ef7da82485a0600cf54f9002b83c0770c2ac86eca0f8779a870\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:39Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.175777 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.175807 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.175816 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.175831 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.175841 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:39Z","lastTransitionTime":"2025-10-11T03:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.187541 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:39Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.200820 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:39Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.213602 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:39Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.228222 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:39Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.244577 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:39Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.277514 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.277553 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.277561 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.277575 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.277585 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:39Z","lastTransitionTime":"2025-10-11T03:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.380924 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.380963 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.380972 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.380988 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.380998 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:39Z","lastTransitionTime":"2025-10-11T03:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.433862 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.433917 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:51:39 crc kubenswrapper[4967]: E1011 03:51:39.434031 4967 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 11 03:51:39 crc kubenswrapper[4967]: E1011 03:51:39.434053 4967 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 11 03:51:39 crc kubenswrapper[4967]: E1011 03:51:39.434146 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-11 03:51:47.434127297 +0000 UTC m=+35.397336230 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 11 03:51:39 crc kubenswrapper[4967]: E1011 03:51:39.434201 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-11 03:51:47.434176928 +0000 UTC m=+35.397385921 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.483933 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.483992 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.484011 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.484036 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.484054 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:39Z","lastTransitionTime":"2025-10-11T03:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.534988 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.535054 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:51:39 crc kubenswrapper[4967]: E1011 03:51:39.535195 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 11 03:51:39 crc kubenswrapper[4967]: E1011 03:51:39.535214 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 11 03:51:39 crc kubenswrapper[4967]: E1011 03:51:39.535227 4967 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:51:39 crc kubenswrapper[4967]: E1011 03:51:39.535275 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-11 03:51:47.535259527 +0000 UTC m=+35.498468470 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:51:39 crc kubenswrapper[4967]: E1011 03:51:39.535622 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 11 03:51:39 crc kubenswrapper[4967]: E1011 03:51:39.535643 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 11 03:51:39 crc kubenswrapper[4967]: E1011 03:51:39.535656 4967 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:51:39 crc kubenswrapper[4967]: E1011 03:51:39.535692 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-11 03:51:47.535682126 +0000 UTC m=+35.498891069 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.586670 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.586728 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.586746 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.586770 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.586786 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:39Z","lastTransitionTime":"2025-10-11T03:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.636162 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:51:39 crc kubenswrapper[4967]: E1011 03:51:39.636357 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:51:47.636326395 +0000 UTC m=+35.599535368 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.690347 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.690399 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.690418 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.690442 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.690461 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:39Z","lastTransitionTime":"2025-10-11T03:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.793402 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.793461 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.793483 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.793507 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.793525 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:39Z","lastTransitionTime":"2025-10-11T03:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.814292 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.814353 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:51:39 crc kubenswrapper[4967]: E1011 03:51:39.814467 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:51:39 crc kubenswrapper[4967]: E1011 03:51:39.814569 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.815701 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:51:39 crc kubenswrapper[4967]: E1011 03:51:39.815862 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.896926 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.896995 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.897017 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.897046 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:39 crc kubenswrapper[4967]: I1011 03:51:39.897067 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:39Z","lastTransitionTime":"2025-10-11T03:51:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.000328 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.000395 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.000419 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.000448 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.000473 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:40Z","lastTransitionTime":"2025-10-11T03:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.026647 4967 generic.go:334] "Generic (PLEG): container finished" podID="f89f20ee-089e-446a-85c4-933aacdb1eba" containerID="e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3" exitCode=0 Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.026719 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" event={"ID":"f89f20ee-089e-446a-85c4-933aacdb1eba","Type":"ContainerDied","Data":"e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3"} Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.083793 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:40Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.098407 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:40Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.103749 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.103786 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.103797 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.103842 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.103857 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:40Z","lastTransitionTime":"2025-10-11T03:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.117880 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:40Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.130004 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:40Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.146906 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:40Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.173531 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:40Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.189952 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:40Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.201864 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:40Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.206800 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.206842 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.206853 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.206868 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.206879 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:40Z","lastTransitionTime":"2025-10-11T03:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.217244 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:40Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.240604 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:40Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.253999 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad406c24472c9ef7da82485a0600cf54f9002b83c0770c2ac86eca0f8779a870\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:40Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.269611 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:40Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.284306 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:40Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.299508 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:40Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.309460 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.309503 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.309513 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.309531 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.309543 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:40Z","lastTransitionTime":"2025-10-11T03:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.317482 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:40Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.412307 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.412360 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.412378 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.412402 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.412418 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:40Z","lastTransitionTime":"2025-10-11T03:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.517769 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.517830 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.517847 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.517871 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.517889 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:40Z","lastTransitionTime":"2025-10-11T03:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.621202 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.621262 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.621281 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.621306 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.621326 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:40Z","lastTransitionTime":"2025-10-11T03:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.724515 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.724568 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.724584 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.724607 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.724624 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:40Z","lastTransitionTime":"2025-10-11T03:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.827702 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.827780 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.827807 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.827830 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.827849 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:40Z","lastTransitionTime":"2025-10-11T03:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.930474 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.930537 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.930557 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.930583 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:40 crc kubenswrapper[4967]: I1011 03:51:40.930599 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:40Z","lastTransitionTime":"2025-10-11T03:51:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.032434 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.032472 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.032486 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.032501 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.032511 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:41Z","lastTransitionTime":"2025-10-11T03:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.034227 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" event={"ID":"270ac8da-b130-4f4f-bbf7-655229b64446","Type":"ContainerStarted","Data":"f02c01578f749f9a0c3565a7c7270bb6927af47e6771226ac8ebbc0e226157bb"} Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.034521 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.034553 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.034572 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.039544 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" event={"ID":"f89f20ee-089e-446a-85c4-933aacdb1eba","Type":"ContainerStarted","Data":"257b2303215c18aa4500478fbfa39f9e1b95678c6d39ea51e4324c58f51dbd30"} Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.055857 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.062596 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.065960 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.076743 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.096392 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.112810 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.135186 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.135244 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.135262 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.135287 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.135308 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:41Z","lastTransitionTime":"2025-10-11T03:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.144233 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f02c01578f749f9a0c3565a7c7270bb6927af47e6771226ac8ebbc0e226157bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.170359 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.185223 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.198938 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.213494 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.228368 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.238235 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.238285 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.238300 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.238322 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.238335 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:41Z","lastTransitionTime":"2025-10-11T03:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.245735 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.257213 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.275838 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.291990 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.303691 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad406c24472c9ef7da82485a0600cf54f9002b83c0770c2ac86eca0f8779a870\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.326606 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f02c01578f749f9a0c3565a7c7270bb6927af47e6771226ac8ebbc0e226157bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.341606 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.341668 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.341684 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.341708 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.341725 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:41Z","lastTransitionTime":"2025-10-11T03:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.346196 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.386943 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.407330 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.428713 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.444966 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.445016 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.445032 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.445055 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.445100 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:41Z","lastTransitionTime":"2025-10-11T03:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.450152 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.470583 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.492417 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.506820 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.530902 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://257b2303215c18aa4500478fbfa39f9e1b95678c6d39ea51e4324c58f51dbd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.548062 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.548177 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.548197 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.548220 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.548237 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:41Z","lastTransitionTime":"2025-10-11T03:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.548422 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad406c24472c9ef7da82485a0600cf54f9002b83c0770c2ac86eca0f8779a870\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.574670 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.596184 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.616500 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.638497 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:41Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.651182 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.651224 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.651239 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.651264 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.651278 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:41Z","lastTransitionTime":"2025-10-11T03:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.755247 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.755316 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.755337 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.755363 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.755381 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:41Z","lastTransitionTime":"2025-10-11T03:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.814314 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.814388 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.814528 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:51:41 crc kubenswrapper[4967]: E1011 03:51:41.814677 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:51:41 crc kubenswrapper[4967]: E1011 03:51:41.814850 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:51:41 crc kubenswrapper[4967]: E1011 03:51:41.815148 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.858878 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.858955 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.858981 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.859013 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.859038 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:41Z","lastTransitionTime":"2025-10-11T03:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.962208 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.962272 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.962289 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.962316 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:41 crc kubenswrapper[4967]: I1011 03:51:41.962338 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:41Z","lastTransitionTime":"2025-10-11T03:51:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.064734 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.064781 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.064795 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.064814 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.064825 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:42Z","lastTransitionTime":"2025-10-11T03:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.168099 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.168154 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.168170 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.168195 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.168213 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:42Z","lastTransitionTime":"2025-10-11T03:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.271021 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.271120 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.271143 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.271178 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.271199 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:42Z","lastTransitionTime":"2025-10-11T03:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.375907 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.375982 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.376000 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.376026 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.376043 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:42Z","lastTransitionTime":"2025-10-11T03:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.479286 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.479356 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.479373 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.479399 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.479417 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:42Z","lastTransitionTime":"2025-10-11T03:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.582452 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.582512 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.582530 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.582552 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.582571 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:42Z","lastTransitionTime":"2025-10-11T03:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.686154 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.686216 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.686233 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.686258 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.686276 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:42Z","lastTransitionTime":"2025-10-11T03:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.790243 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.790313 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.790329 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.790356 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.790381 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:42Z","lastTransitionTime":"2025-10-11T03:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.838469 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.860319 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.880209 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.893789 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.893839 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.893851 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.893865 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.893875 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:42Z","lastTransitionTime":"2025-10-11T03:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.898509 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.925524 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f02c01578f749f9a0c3565a7c7270bb6927af47e6771226ac8ebbc0e226157bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.941244 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.951145 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.961562 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.978322 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.988625 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.996128 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.996166 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.996177 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.996198 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.996209 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:42Z","lastTransitionTime":"2025-10-11T03:51:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:42 crc kubenswrapper[4967]: I1011 03:51:42.999553 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.011894 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.024232 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.037597 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://257b2303215c18aa4500478fbfa39f9e1b95678c6d39ea51e4324c58f51dbd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.047199 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad406c24472c9ef7da82485a0600cf54f9002b83c0770c2ac86eca0f8779a870\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.098545 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.098580 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.098592 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.098607 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.098621 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:43Z","lastTransitionTime":"2025-10-11T03:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.201623 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.201653 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.201661 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.201675 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.201684 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:43Z","lastTransitionTime":"2025-10-11T03:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.305197 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.305246 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.305259 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.305277 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.305288 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:43Z","lastTransitionTime":"2025-10-11T03:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.408470 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.408530 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.408546 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.408569 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.408587 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:43Z","lastTransitionTime":"2025-10-11T03:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.511941 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.512000 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.512022 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.512053 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.512105 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:43Z","lastTransitionTime":"2025-10-11T03:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.615381 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.615431 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.615449 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.615474 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.615490 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:43Z","lastTransitionTime":"2025-10-11T03:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.719406 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.719466 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.719482 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.719506 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.719525 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:43Z","lastTransitionTime":"2025-10-11T03:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.814754 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.814829 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.814872 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:51:43 crc kubenswrapper[4967]: E1011 03:51:43.814952 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:51:43 crc kubenswrapper[4967]: E1011 03:51:43.815159 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:51:43 crc kubenswrapper[4967]: E1011 03:51:43.815296 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.823041 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.823135 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.823154 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.823202 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.823220 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:43Z","lastTransitionTime":"2025-10-11T03:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.926974 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.927026 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.927043 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.927066 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:43 crc kubenswrapper[4967]: I1011 03:51:43.927109 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:43Z","lastTransitionTime":"2025-10-11T03:51:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.030044 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.030184 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.030217 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.030245 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.030262 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:44Z","lastTransitionTime":"2025-10-11T03:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.053529 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lw6dx_270ac8da-b130-4f4f-bbf7-655229b64446/ovnkube-controller/0.log" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.058140 4967 generic.go:334] "Generic (PLEG): container finished" podID="270ac8da-b130-4f4f-bbf7-655229b64446" containerID="f02c01578f749f9a0c3565a7c7270bb6927af47e6771226ac8ebbc0e226157bb" exitCode=1 Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.058219 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" event={"ID":"270ac8da-b130-4f4f-bbf7-655229b64446","Type":"ContainerDied","Data":"f02c01578f749f9a0c3565a7c7270bb6927af47e6771226ac8ebbc0e226157bb"} Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.059228 4967 scope.go:117] "RemoveContainer" containerID="f02c01578f749f9a0c3565a7c7270bb6927af47e6771226ac8ebbc0e226157bb" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.091882 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://f02c01578f749f9a0c3565a7c7270bb6927af47e6771226ac8ebbc0e226157bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f02c01578f749f9a0c3565a7c7270bb6927af47e6771226ac8ebbc0e226157bb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:51:43Z\\\",\\\"message\\\":\\\"github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:51:43.446126 6277 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:51:43.446220 6277 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:51:43.446453 6277 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:51:43.446612 6277 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1011 03:51:43.446713 6277 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:51:43.446656 6277 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:44Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.114534 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:44Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.132342 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.132388 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.132404 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.132426 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.132440 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:44Z","lastTransitionTime":"2025-10-11T03:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.138599 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:44Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.151437 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:44Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.168441 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:44Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.182176 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:44Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.202021 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:44Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.219180 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:44Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.231444 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:44Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.235238 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.235311 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.235336 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.235368 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.235391 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:44Z","lastTransitionTime":"2025-10-11T03:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.252395 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://257b2303215c18aa4500478fbfa39f9e1b95678c6d39ea51e4324c58f51dbd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:44Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.265172 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad406c24472c9ef7da82485a0600cf54f9002b83c0770c2ac86eca0f8779a870\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:44Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.281183 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:44Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.296823 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:44Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.309034 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:44Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.323472 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:44Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.338408 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.338442 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.338453 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.338469 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.338482 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:44Z","lastTransitionTime":"2025-10-11T03:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.441255 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.441307 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.441318 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.441337 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.441351 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:44Z","lastTransitionTime":"2025-10-11T03:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.543867 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.543930 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.543953 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.543983 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.544006 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:44Z","lastTransitionTime":"2025-10-11T03:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.646191 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.646236 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.646245 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.646260 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.646269 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:44Z","lastTransitionTime":"2025-10-11T03:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.748717 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.748781 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.748800 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.748827 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.748845 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:44Z","lastTransitionTime":"2025-10-11T03:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.851327 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.851373 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.851386 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.851403 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.851415 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:44Z","lastTransitionTime":"2025-10-11T03:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.954185 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.954265 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.954287 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.954320 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:44 crc kubenswrapper[4967]: I1011 03:51:44.954342 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:44Z","lastTransitionTime":"2025-10-11T03:51:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.056155 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.056191 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.056200 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.056213 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.056222 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:45Z","lastTransitionTime":"2025-10-11T03:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.062452 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lw6dx_270ac8da-b130-4f4f-bbf7-655229b64446/ovnkube-controller/0.log" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.065445 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" event={"ID":"270ac8da-b130-4f4f-bbf7-655229b64446","Type":"ContainerStarted","Data":"1c040b1dc41b86ac58bb6fe7d65c41fcac12bfcdf706e42a8cbf3389f90c09f3"} Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.065815 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.086814 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:45Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.104685 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:45Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.115935 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:45Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.126121 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:45Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.151827 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c040b1dc41b86ac58bb6fe7d65c41fcac12bfcdf706e42a8cbf3389f90c09f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f02c01578f749f9a0c3565a7c7270bb6927af47e6771226ac8ebbc0e226157bb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:51:43Z\\\",\\\"message\\\":\\\"github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:51:43.446126 6277 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:51:43.446220 6277 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:51:43.446453 6277 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:51:43.446612 6277 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1011 03:51:43.446713 6277 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:51:43.446656 6277 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:45Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.158786 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.158822 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.158833 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.158849 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.158863 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:45Z","lastTransitionTime":"2025-10-11T03:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.165977 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:45Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.185637 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:45Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.201128 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:45Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.218858 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:45Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.254568 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:45Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.261598 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.261671 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.261699 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.261732 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.261756 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:45Z","lastTransitionTime":"2025-10-11T03:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.276112 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:45Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.296796 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:45Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.318858 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:45Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.343149 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://257b2303215c18aa4500478fbfa39f9e1b95678c6d39ea51e4324c58f51dbd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:45Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.357636 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad406c24472c9ef7da82485a0600cf54f9002b83c0770c2ac86eca0f8779a870\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:45Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.364715 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.364767 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.364787 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.364815 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.364843 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:45Z","lastTransitionTime":"2025-10-11T03:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.467660 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.467713 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.467722 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.467735 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.467744 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:45Z","lastTransitionTime":"2025-10-11T03:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.570689 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.570766 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.570786 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.570812 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.570830 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:45Z","lastTransitionTime":"2025-10-11T03:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.673834 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.673889 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.673909 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.673931 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.673947 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:45Z","lastTransitionTime":"2025-10-11T03:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.776774 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.776893 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.776920 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.776955 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.776991 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:45Z","lastTransitionTime":"2025-10-11T03:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.814105 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.814164 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.814118 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:51:45 crc kubenswrapper[4967]: E1011 03:51:45.814312 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:51:45 crc kubenswrapper[4967]: E1011 03:51:45.814400 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:51:45 crc kubenswrapper[4967]: E1011 03:51:45.814544 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.880377 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.880462 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.880479 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.880501 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.880516 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:45Z","lastTransitionTime":"2025-10-11T03:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.983557 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.983615 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.983632 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.983656 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:45 crc kubenswrapper[4967]: I1011 03:51:45.983673 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:45Z","lastTransitionTime":"2025-10-11T03:51:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.071462 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lw6dx_270ac8da-b130-4f4f-bbf7-655229b64446/ovnkube-controller/1.log" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.072349 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lw6dx_270ac8da-b130-4f4f-bbf7-655229b64446/ovnkube-controller/0.log" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.076645 4967 generic.go:334] "Generic (PLEG): container finished" podID="270ac8da-b130-4f4f-bbf7-655229b64446" containerID="1c040b1dc41b86ac58bb6fe7d65c41fcac12bfcdf706e42a8cbf3389f90c09f3" exitCode=1 Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.076702 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" event={"ID":"270ac8da-b130-4f4f-bbf7-655229b64446","Type":"ContainerDied","Data":"1c040b1dc41b86ac58bb6fe7d65c41fcac12bfcdf706e42a8cbf3389f90c09f3"} Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.076776 4967 scope.go:117] "RemoveContainer" containerID="f02c01578f749f9a0c3565a7c7270bb6927af47e6771226ac8ebbc0e226157bb" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.077751 4967 scope.go:117] "RemoveContainer" containerID="1c040b1dc41b86ac58bb6fe7d65c41fcac12bfcdf706e42a8cbf3389f90c09f3" Oct 11 03:51:46 crc kubenswrapper[4967]: E1011 03:51:46.078004 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-lw6dx_openshift-ovn-kubernetes(270ac8da-b130-4f4f-bbf7-655229b64446)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.086449 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.086518 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.086542 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.086571 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.086597 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:46Z","lastTransitionTime":"2025-10-11T03:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.098411 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:46Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.120854 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:46Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.139150 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz"] Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.139825 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.140048 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:46Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.141810 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.142562 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.160833 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:46Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.185591 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c040b1dc41b86ac58bb6fe7d65c41fcac12bfcdf706e42a8cbf3389f90c09f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f02c01578f749f9a0c3565a7c7270bb6927af47e6771226ac8ebbc0e226157bb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:51:43Z\\\",\\\"message\\\":\\\"github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:51:43.446126 6277 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:51:43.446220 6277 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:51:43.446453 6277 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:51:43.446612 6277 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1011 03:51:43.446713 6277 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:51:43.446656 6277 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c040b1dc41b86ac58bb6fe7d65c41fcac12bfcdf706e42a8cbf3389f90c09f3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:51:45Z\\\",\\\"message\\\":\\\"start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:45Z is after 2025-08-24T17:21:41Z]\\\\nI1011 03:51:45.040782 6396 services_controller.go:434] Service openshift-console-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-console-operator e4559ce3-2d5a-470f-b8bf-4c8b054d2335 11843 0 2025-02-23 05:38:55 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:console-operator] map[capability.openshift.io/name:Console include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc006efe87f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:46Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.189375 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.189433 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.189451 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.189476 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.189495 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:46Z","lastTransitionTime":"2025-10-11T03:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.205704 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:46Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.210686 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/844146aa-36ac-4bee-8015-0d9850d94c4d-env-overrides\") pod \"ovnkube-control-plane-749d76644c-hs9vz\" (UID: \"844146aa-36ac-4bee-8015-0d9850d94c4d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.210774 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5jll\" (UniqueName: \"kubernetes.io/projected/844146aa-36ac-4bee-8015-0d9850d94c4d-kube-api-access-r5jll\") pod \"ovnkube-control-plane-749d76644c-hs9vz\" (UID: \"844146aa-36ac-4bee-8015-0d9850d94c4d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.210798 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/844146aa-36ac-4bee-8015-0d9850d94c4d-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-hs9vz\" (UID: \"844146aa-36ac-4bee-8015-0d9850d94c4d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.211160 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/844146aa-36ac-4bee-8015-0d9850d94c4d-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-hs9vz\" (UID: \"844146aa-36ac-4bee-8015-0d9850d94c4d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.232629 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:46Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.252265 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:46Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.265811 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:46Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.282339 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:46Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.292263 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.292304 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.292316 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.292335 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.292349 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:46Z","lastTransitionTime":"2025-10-11T03:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.312105 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/844146aa-36ac-4bee-8015-0d9850d94c4d-env-overrides\") pod \"ovnkube-control-plane-749d76644c-hs9vz\" (UID: \"844146aa-36ac-4bee-8015-0d9850d94c4d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.312214 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5jll\" (UniqueName: \"kubernetes.io/projected/844146aa-36ac-4bee-8015-0d9850d94c4d-kube-api-access-r5jll\") pod \"ovnkube-control-plane-749d76644c-hs9vz\" (UID: \"844146aa-36ac-4bee-8015-0d9850d94c4d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.312254 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/844146aa-36ac-4bee-8015-0d9850d94c4d-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-hs9vz\" (UID: \"844146aa-36ac-4bee-8015-0d9850d94c4d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.312903 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/844146aa-36ac-4bee-8015-0d9850d94c4d-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-hs9vz\" (UID: \"844146aa-36ac-4bee-8015-0d9850d94c4d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.312930 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/844146aa-36ac-4bee-8015-0d9850d94c4d-env-overrides\") pod \"ovnkube-control-plane-749d76644c-hs9vz\" (UID: \"844146aa-36ac-4bee-8015-0d9850d94c4d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.313560 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/844146aa-36ac-4bee-8015-0d9850d94c4d-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-hs9vz\" (UID: \"844146aa-36ac-4bee-8015-0d9850d94c4d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.314858 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:46Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.321678 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/844146aa-36ac-4bee-8015-0d9850d94c4d-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-hs9vz\" (UID: \"844146aa-36ac-4bee-8015-0d9850d94c4d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.334688 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:46Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.344188 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5jll\" (UniqueName: \"kubernetes.io/projected/844146aa-36ac-4bee-8015-0d9850d94c4d-kube-api-access-r5jll\") pod \"ovnkube-control-plane-749d76644c-hs9vz\" (UID: \"844146aa-36ac-4bee-8015-0d9850d94c4d\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.353971 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:46Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.375729 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://257b2303215c18aa4500478fbfa39f9e1b95678c6d39ea51e4324c58f51dbd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:46Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.392374 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad406c24472c9ef7da82485a0600cf54f9002b83c0770c2ac86eca0f8779a870\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:46Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.394823 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.395125 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.395349 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.395561 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.395768 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:46Z","lastTransitionTime":"2025-10-11T03:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.416178 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:46Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.435527 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:46Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.455506 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:46Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.458454 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.469898 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:46Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:46 crc kubenswrapper[4967]: W1011 03:51:46.480710 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod844146aa_36ac_4bee_8015_0d9850d94c4d.slice/crio-0de0cb3110e1aae0eb0a6194541dd54103e8d1e50cde83ce78785ef8066f2852 WatchSource:0}: Error finding container 0de0cb3110e1aae0eb0a6194541dd54103e8d1e50cde83ce78785ef8066f2852: Status 404 returned error can't find the container with id 0de0cb3110e1aae0eb0a6194541dd54103e8d1e50cde83ce78785ef8066f2852 Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.497789 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.497828 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.497839 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.497856 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.497869 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:46Z","lastTransitionTime":"2025-10-11T03:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.514181 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c040b1dc41b86ac58bb6fe7d65c41fcac12bfcdf706e42a8cbf3389f90c09f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f02c01578f749f9a0c3565a7c7270bb6927af47e6771226ac8ebbc0e226157bb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:51:43Z\\\",\\\"message\\\":\\\"github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:51:43.446126 6277 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:51:43.446220 6277 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:51:43.446453 6277 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:51:43.446612 6277 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1011 03:51:43.446713 6277 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:51:43.446656 6277 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c040b1dc41b86ac58bb6fe7d65c41fcac12bfcdf706e42a8cbf3389f90c09f3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:51:45Z\\\",\\\"message\\\":\\\"start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:45Z is after 2025-08-24T17:21:41Z]\\\\nI1011 03:51:45.040782 6396 services_controller.go:434] Service openshift-console-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-console-operator e4559ce3-2d5a-470f-b8bf-4c8b054d2335 11843 0 2025-02-23 05:38:55 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:console-operator] map[capability.openshift.io/name:Console include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc006efe87f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:46Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.540009 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:46Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.558750 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:46Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.572327 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:46Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.591644 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:46Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.601854 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.601955 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.601981 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.602010 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.602031 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:46Z","lastTransitionTime":"2025-10-11T03:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.622381 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:46Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.635993 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:46Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.651691 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:46Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.667027 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:46Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.688851 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://257b2303215c18aa4500478fbfa39f9e1b95678c6d39ea51e4324c58f51dbd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:46Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.701789 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad406c24472c9ef7da82485a0600cf54f9002b83c0770c2ac86eca0f8779a870\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:46Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.705475 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.705513 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.705526 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.705543 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.705554 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:46Z","lastTransitionTime":"2025-10-11T03:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.719207 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"844146aa-36ac-4bee-8015-0d9850d94c4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:46Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:46Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hs9vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:46Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.809015 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.809108 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.809136 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.809166 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.809193 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:46Z","lastTransitionTime":"2025-10-11T03:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.912944 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.913005 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.913018 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.913037 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:46 crc kubenswrapper[4967]: I1011 03:51:46.913054 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:46Z","lastTransitionTime":"2025-10-11T03:51:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.015388 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.015440 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.015455 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.015473 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.015486 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:47Z","lastTransitionTime":"2025-10-11T03:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.082161 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" event={"ID":"844146aa-36ac-4bee-8015-0d9850d94c4d","Type":"ContainerStarted","Data":"ef8f8a8bafbf88541375edb7c1b2134dec31a0d5cd46462a3271d24131bed4c9"} Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.082204 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" event={"ID":"844146aa-36ac-4bee-8015-0d9850d94c4d","Type":"ContainerStarted","Data":"3cafb41d365e674a55638718df4e540ff77b6bd8e0d2d94bdb480b6617c0942f"} Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.082213 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" event={"ID":"844146aa-36ac-4bee-8015-0d9850d94c4d","Type":"ContainerStarted","Data":"0de0cb3110e1aae0eb0a6194541dd54103e8d1e50cde83ce78785ef8066f2852"} Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.084583 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lw6dx_270ac8da-b130-4f4f-bbf7-655229b64446/ovnkube-controller/1.log" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.088181 4967 scope.go:117] "RemoveContainer" containerID="1c040b1dc41b86ac58bb6fe7d65c41fcac12bfcdf706e42a8cbf3389f90c09f3" Oct 11 03:51:47 crc kubenswrapper[4967]: E1011 03:51:47.088312 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-lw6dx_openshift-ovn-kubernetes(270ac8da-b130-4f4f-bbf7-655229b64446)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.116236 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c040b1dc41b86ac58bb6fe7d65c41fcac12bfcdf706e42a8cbf3389f90c09f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f02c01578f749f9a0c3565a7c7270bb6927af47e6771226ac8ebbc0e226157bb\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:51:43Z\\\",\\\"message\\\":\\\"github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:51:43.446126 6277 reflector.go:311] Stopping reflector *v1.EgressQoS (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressqos/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:51:43.446220 6277 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:51:43.446453 6277 reflector.go:311] Stopping reflector *v1.EgressIP (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressip/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:51:43.446612 6277 reflector.go:311] Stopping reflector *v1alpha1.AdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/factory.go:141\\\\nI1011 03:51:43.446713 6277 reflector.go:311] Stopping reflector *v1.ClusterUserDefinedNetwork (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/userdefinednetwork/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:51:43.446656 6277 reflector.go:311] Stopping reflector *v1alpha1.BaselineAdminNetworkPolicy (0s) from sigs.k8s.io/network-policy-api/pkg/client/informers/externalversions/f\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c040b1dc41b86ac58bb6fe7d65c41fcac12bfcdf706e42a8cbf3389f90c09f3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:51:45Z\\\",\\\"message\\\":\\\"start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:45Z is after 2025-08-24T17:21:41Z]\\\\nI1011 03:51:45.040782 6396 services_controller.go:434] Service openshift-console-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-console-operator e4559ce3-2d5a-470f-b8bf-4c8b054d2335 11843 0 2025-02-23 05:38:55 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:console-operator] map[capability.openshift.io/name:Console include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc006efe87f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:47Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.118197 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.118229 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.118242 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.118258 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.118271 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:47Z","lastTransitionTime":"2025-10-11T03:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.133837 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:47Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.157703 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:47Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.174029 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:47Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.190955 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:47Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.210718 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:47Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.219928 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.219969 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.219982 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.220003 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.220015 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:47Z","lastTransitionTime":"2025-10-11T03:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.225556 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:47Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.238798 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:47Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.249049 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:47Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.264107 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://257b2303215c18aa4500478fbfa39f9e1b95678c6d39ea51e4324c58f51dbd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:47Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.273972 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad406c24472c9ef7da82485a0600cf54f9002b83c0770c2ac86eca0f8779a870\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:47Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.285739 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"844146aa-36ac-4bee-8015-0d9850d94c4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cafb41d365e674a55638718df4e540ff77b6bd8e0d2d94bdb480b6617c0942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8f8a8bafbf88541375edb7c1b2134dec31a0d5cd46462a3271d24131bed4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hs9vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:47Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.301383 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:47Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.313455 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:47Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.322122 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.322161 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.322172 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.322186 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.322195 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:47Z","lastTransitionTime":"2025-10-11T03:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.328750 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:47Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.345295 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:47Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.359742 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:47Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.371724 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:47Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.388231 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:47Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.399187 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:47Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.423484 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c040b1dc41b86ac58bb6fe7d65c41fcac12bfcdf706e42a8cbf3389f90c09f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c040b1dc41b86ac58bb6fe7d65c41fcac12bfcdf706e42a8cbf3389f90c09f3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:51:45Z\\\",\\\"message\\\":\\\"start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:45Z is after 2025-08-24T17:21:41Z]\\\\nI1011 03:51:45.040782 6396 services_controller.go:434] Service openshift-console-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-console-operator e4559ce3-2d5a-470f-b8bf-4c8b054d2335 11843 0 2025-02-23 05:38:55 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:console-operator] map[capability.openshift.io/name:Console include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc006efe87f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-lw6dx_openshift-ovn-kubernetes(270ac8da-b130-4f4f-bbf7-655229b64446)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:47Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.425180 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.425208 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.425216 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.425232 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.425240 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:47Z","lastTransitionTime":"2025-10-11T03:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.434451 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:47Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.450638 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:47Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.474641 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:47Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.491210 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:47Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.506216 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:47Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.522470 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:47Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.523029 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:51:47 crc kubenswrapper[4967]: E1011 03:51:47.523201 4967 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 11 03:51:47 crc kubenswrapper[4967]: E1011 03:51:47.523283 4967 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.523203 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:51:47 crc kubenswrapper[4967]: E1011 03:51:47.523288 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-11 03:52:03.523271477 +0000 UTC m=+51.486480410 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 11 03:51:47 crc kubenswrapper[4967]: E1011 03:51:47.523375 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-11 03:52:03.523354179 +0000 UTC m=+51.486563152 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.527446 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.527508 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.527526 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.527551 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.527570 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:47Z","lastTransitionTime":"2025-10-11T03:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.537330 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:47Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.549756 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:47Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.563964 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://257b2303215c18aa4500478fbfa39f9e1b95678c6d39ea51e4324c58f51dbd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:47Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.575927 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad406c24472c9ef7da82485a0600cf54f9002b83c0770c2ac86eca0f8779a870\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:47Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.587331 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"844146aa-36ac-4bee-8015-0d9850d94c4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cafb41d365e674a55638718df4e540ff77b6bd8e0d2d94bdb480b6617c0942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8f8a8bafbf88541375edb7c1b2134dec31a0d5cd46462a3271d24131bed4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hs9vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:47Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.624364 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.624476 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:51:47 crc kubenswrapper[4967]: E1011 03:51:47.624602 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 11 03:51:47 crc kubenswrapper[4967]: E1011 03:51:47.624624 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 11 03:51:47 crc kubenswrapper[4967]: E1011 03:51:47.624649 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 11 03:51:47 crc kubenswrapper[4967]: E1011 03:51:47.624657 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 11 03:51:47 crc kubenswrapper[4967]: E1011 03:51:47.624676 4967 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:51:47 crc kubenswrapper[4967]: E1011 03:51:47.624677 4967 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:51:47 crc kubenswrapper[4967]: E1011 03:51:47.624741 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-11 03:52:03.624721514 +0000 UTC m=+51.587930467 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:51:47 crc kubenswrapper[4967]: E1011 03:51:47.624773 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-11 03:52:03.624760455 +0000 UTC m=+51.587969638 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.630476 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.630527 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.630543 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.630566 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.630583 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:47Z","lastTransitionTime":"2025-10-11T03:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.725709 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:51:47 crc kubenswrapper[4967]: E1011 03:51:47.725956 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:52:03.725917505 +0000 UTC m=+51.689126468 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.733922 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.733964 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.733976 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.733996 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.734009 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:47Z","lastTransitionTime":"2025-10-11T03:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.814815 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.814929 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.814855 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:51:47 crc kubenswrapper[4967]: E1011 03:51:47.815021 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:51:47 crc kubenswrapper[4967]: E1011 03:51:47.815157 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:51:47 crc kubenswrapper[4967]: E1011 03:51:47.815252 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.837782 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.837844 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.837860 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.837884 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.837901 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:47Z","lastTransitionTime":"2025-10-11T03:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.941177 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.941264 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.941280 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.941306 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.941323 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:47Z","lastTransitionTime":"2025-10-11T03:51:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.993591 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-p24x8"] Oct 11 03:51:47 crc kubenswrapper[4967]: I1011 03:51:47.994053 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:51:47 crc kubenswrapper[4967]: E1011 03:51:47.994127 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.011601 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:48Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.039663 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:48Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.043928 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.043959 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.043969 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.043984 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.043998 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:48Z","lastTransitionTime":"2025-10-11T03:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.055187 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:48Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.071479 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:48Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.088739 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:48Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.106947 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:48Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.124518 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:48Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.131047 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d22c953c-0315-406e-9c5d-f1759837750e-metrics-certs\") pod \"network-metrics-daemon-p24x8\" (UID: \"d22c953c-0315-406e-9c5d-f1759837750e\") " pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.131182 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6glk9\" (UniqueName: \"kubernetes.io/projected/d22c953c-0315-406e-9c5d-f1759837750e-kube-api-access-6glk9\") pod \"network-metrics-daemon-p24x8\" (UID: \"d22c953c-0315-406e-9c5d-f1759837750e\") " pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.139264 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:48Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.145695 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.145754 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.145770 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.145795 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.145812 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:48Z","lastTransitionTime":"2025-10-11T03:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.163357 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://257b2303215c18aa4500478fbfa39f9e1b95678c6d39ea51e4324c58f51dbd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:48Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.176872 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad406c24472c9ef7da82485a0600cf54f9002b83c0770c2ac86eca0f8779a870\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:48Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.198598 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"844146aa-36ac-4bee-8015-0d9850d94c4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cafb41d365e674a55638718df4e540ff77b6bd8e0d2d94bdb480b6617c0942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8f8a8bafbf88541375edb7c1b2134dec31a0d5cd46462a3271d24131bed4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hs9vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:48Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.219384 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:48Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.232809 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d22c953c-0315-406e-9c5d-f1759837750e-metrics-certs\") pod \"network-metrics-daemon-p24x8\" (UID: \"d22c953c-0315-406e-9c5d-f1759837750e\") " pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.232937 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6glk9\" (UniqueName: \"kubernetes.io/projected/d22c953c-0315-406e-9c5d-f1759837750e-kube-api-access-6glk9\") pod \"network-metrics-daemon-p24x8\" (UID: \"d22c953c-0315-406e-9c5d-f1759837750e\") " pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:51:48 crc kubenswrapper[4967]: E1011 03:51:48.233012 4967 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 11 03:51:48 crc kubenswrapper[4967]: E1011 03:51:48.233114 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d22c953c-0315-406e-9c5d-f1759837750e-metrics-certs podName:d22c953c-0315-406e-9c5d-f1759837750e nodeName:}" failed. No retries permitted until 2025-10-11 03:51:48.733063665 +0000 UTC m=+36.696272638 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d22c953c-0315-406e-9c5d-f1759837750e-metrics-certs") pod "network-metrics-daemon-p24x8" (UID: "d22c953c-0315-406e-9c5d-f1759837750e") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.241546 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:48Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.248619 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.248706 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.248733 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.248761 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.248783 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:48Z","lastTransitionTime":"2025-10-11T03:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.261109 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:48Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.262547 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6glk9\" (UniqueName: \"kubernetes.io/projected/d22c953c-0315-406e-9c5d-f1759837750e-kube-api-access-6glk9\") pod \"network-metrics-daemon-p24x8\" (UID: \"d22c953c-0315-406e-9c5d-f1759837750e\") " pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.280412 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:48Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.299731 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.299897 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.299928 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.299963 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.299989 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:48Z","lastTransitionTime":"2025-10-11T03:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.313646 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c040b1dc41b86ac58bb6fe7d65c41fcac12bfcdf706e42a8cbf3389f90c09f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c040b1dc41b86ac58bb6fe7d65c41fcac12bfcdf706e42a8cbf3389f90c09f3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:51:45Z\\\",\\\"message\\\":\\\"start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:45Z is after 2025-08-24T17:21:41Z]\\\\nI1011 03:51:45.040782 6396 services_controller.go:434] Service openshift-console-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-console-operator e4559ce3-2d5a-470f-b8bf-4c8b054d2335 11843 0 2025-02-23 05:38:55 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:console-operator] map[capability.openshift.io/name:Console include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc006efe87f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-lw6dx_openshift-ovn-kubernetes(270ac8da-b130-4f4f-bbf7-655229b64446)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:48Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:48 crc kubenswrapper[4967]: E1011 03:51:48.322806 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:48Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.328223 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.328300 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.328328 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.328358 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.328382 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:48Z","lastTransitionTime":"2025-10-11T03:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.334115 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p24x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22c953c-0315-406e-9c5d-f1759837750e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p24x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:48Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:48 crc kubenswrapper[4967]: E1011 03:51:48.349142 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:48Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.354594 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.354676 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.354705 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.354735 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.354754 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:48Z","lastTransitionTime":"2025-10-11T03:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:48 crc kubenswrapper[4967]: E1011 03:51:48.375460 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:48Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.380585 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.380667 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.380691 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.380723 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.380745 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:48Z","lastTransitionTime":"2025-10-11T03:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:48 crc kubenswrapper[4967]: E1011 03:51:48.400852 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:48Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.405597 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.405661 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.405687 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.405721 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.405746 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:48Z","lastTransitionTime":"2025-10-11T03:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:48 crc kubenswrapper[4967]: E1011 03:51:48.423909 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:48Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:48Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:48 crc kubenswrapper[4967]: E1011 03:51:48.424182 4967 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.426390 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.426450 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.426471 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.426496 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.426514 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:48Z","lastTransitionTime":"2025-10-11T03:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.529789 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.529861 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.529882 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.529925 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.529943 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:48Z","lastTransitionTime":"2025-10-11T03:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.632694 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.632783 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.632805 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.632836 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.632856 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:48Z","lastTransitionTime":"2025-10-11T03:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.736122 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.736206 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.736228 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.736256 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.736275 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:48Z","lastTransitionTime":"2025-10-11T03:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.738899 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d22c953c-0315-406e-9c5d-f1759837750e-metrics-certs\") pod \"network-metrics-daemon-p24x8\" (UID: \"d22c953c-0315-406e-9c5d-f1759837750e\") " pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:51:48 crc kubenswrapper[4967]: E1011 03:51:48.739063 4967 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 11 03:51:48 crc kubenswrapper[4967]: E1011 03:51:48.739196 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d22c953c-0315-406e-9c5d-f1759837750e-metrics-certs podName:d22c953c-0315-406e-9c5d-f1759837750e nodeName:}" failed. No retries permitted until 2025-10-11 03:51:49.739170054 +0000 UTC m=+37.702379027 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d22c953c-0315-406e-9c5d-f1759837750e-metrics-certs") pod "network-metrics-daemon-p24x8" (UID: "d22c953c-0315-406e-9c5d-f1759837750e") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.839496 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.839560 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.839578 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.839602 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.839620 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:48Z","lastTransitionTime":"2025-10-11T03:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.942435 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.942491 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.942542 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.942575 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:48 crc kubenswrapper[4967]: I1011 03:51:48.942597 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:48Z","lastTransitionTime":"2025-10-11T03:51:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.045838 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.045920 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.045938 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.045960 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.045977 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:49Z","lastTransitionTime":"2025-10-11T03:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.152985 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.153054 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.153107 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.153151 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.153175 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:49Z","lastTransitionTime":"2025-10-11T03:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.256796 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.256877 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.256894 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.256920 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.256937 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:49Z","lastTransitionTime":"2025-10-11T03:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.359681 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.359766 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.359788 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.359888 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.359914 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:49Z","lastTransitionTime":"2025-10-11T03:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.462969 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.463034 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.463051 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.463113 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.463151 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:49Z","lastTransitionTime":"2025-10-11T03:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.566031 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.566124 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.566142 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.566165 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.566184 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:49Z","lastTransitionTime":"2025-10-11T03:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.670136 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.670210 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.670229 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.670261 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.670283 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:49Z","lastTransitionTime":"2025-10-11T03:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.749531 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d22c953c-0315-406e-9c5d-f1759837750e-metrics-certs\") pod \"network-metrics-daemon-p24x8\" (UID: \"d22c953c-0315-406e-9c5d-f1759837750e\") " pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:51:49 crc kubenswrapper[4967]: E1011 03:51:49.749935 4967 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 11 03:51:49 crc kubenswrapper[4967]: E1011 03:51:49.750168 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d22c953c-0315-406e-9c5d-f1759837750e-metrics-certs podName:d22c953c-0315-406e-9c5d-f1759837750e nodeName:}" failed. No retries permitted until 2025-10-11 03:51:51.750130143 +0000 UTC m=+39.713339146 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d22c953c-0315-406e-9c5d-f1759837750e-metrics-certs") pod "network-metrics-daemon-p24x8" (UID: "d22c953c-0315-406e-9c5d-f1759837750e") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.774239 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.774296 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.774313 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.774339 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.774361 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:49Z","lastTransitionTime":"2025-10-11T03:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.814709 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.814741 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.814867 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:51:49 crc kubenswrapper[4967]: E1011 03:51:49.815123 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.815511 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:51:49 crc kubenswrapper[4967]: E1011 03:51:49.815681 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:51:49 crc kubenswrapper[4967]: E1011 03:51:49.815785 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:51:49 crc kubenswrapper[4967]: E1011 03:51:49.815886 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.877797 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.877894 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.877916 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.877941 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.877961 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:49Z","lastTransitionTime":"2025-10-11T03:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.981423 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.981476 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.981496 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.981522 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:49 crc kubenswrapper[4967]: I1011 03:51:49.981539 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:49Z","lastTransitionTime":"2025-10-11T03:51:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.084016 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.084108 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.084126 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.084149 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.084167 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:50Z","lastTransitionTime":"2025-10-11T03:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.186671 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.186721 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.186740 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.186766 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.186785 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:50Z","lastTransitionTime":"2025-10-11T03:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.289488 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.289534 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.289552 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.289575 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.289595 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:50Z","lastTransitionTime":"2025-10-11T03:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.392971 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.393031 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.393054 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.393123 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.393149 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:50Z","lastTransitionTime":"2025-10-11T03:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.496476 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.496552 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.496575 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.496603 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.496627 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:50Z","lastTransitionTime":"2025-10-11T03:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.599626 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.599716 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.599742 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.599771 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.599794 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:50Z","lastTransitionTime":"2025-10-11T03:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.703163 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.703236 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.703259 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.703288 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.703309 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:50Z","lastTransitionTime":"2025-10-11T03:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.806116 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.806176 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.806193 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.806268 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.806292 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:50Z","lastTransitionTime":"2025-10-11T03:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.909766 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.909843 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.909868 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.909898 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:50 crc kubenswrapper[4967]: I1011 03:51:50.909919 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:50Z","lastTransitionTime":"2025-10-11T03:51:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.013619 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.013667 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.013683 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.013710 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.013731 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:51Z","lastTransitionTime":"2025-10-11T03:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.116887 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.116943 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.116961 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.116985 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.117002 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:51Z","lastTransitionTime":"2025-10-11T03:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.220916 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.220972 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.220996 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.221028 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.221051 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:51Z","lastTransitionTime":"2025-10-11T03:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.255528 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.274542 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:51Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.298690 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:51Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.324653 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.324717 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.324740 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.324769 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.324790 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:51Z","lastTransitionTime":"2025-10-11T03:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.333333 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:51Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.361836 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:51Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.387504 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c040b1dc41b86ac58bb6fe7d65c41fcac12bfcdf706e42a8cbf3389f90c09f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c040b1dc41b86ac58bb6fe7d65c41fcac12bfcdf706e42a8cbf3389f90c09f3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:51:45Z\\\",\\\"message\\\":\\\"start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:45Z is after 2025-08-24T17:21:41Z]\\\\nI1011 03:51:45.040782 6396 services_controller.go:434] Service openshift-console-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-console-operator e4559ce3-2d5a-470f-b8bf-4c8b054d2335 11843 0 2025-02-23 05:38:55 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:console-operator] map[capability.openshift.io/name:Console include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc006efe87f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-lw6dx_openshift-ovn-kubernetes(270ac8da-b130-4f4f-bbf7-655229b64446)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:51Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.397645 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p24x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22c953c-0315-406e-9c5d-f1759837750e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p24x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:51Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.413928 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:51Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.428657 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.428755 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.428930 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.429090 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.429178 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:51Z","lastTransitionTime":"2025-10-11T03:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.447162 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:51Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.464647 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:51Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.482328 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:51Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.501206 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:51Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.519300 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:51Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.532068 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.532181 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.532198 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.532224 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.532244 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:51Z","lastTransitionTime":"2025-10-11T03:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.537262 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:51Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.554795 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:51Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.572874 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://257b2303215c18aa4500478fbfa39f9e1b95678c6d39ea51e4324c58f51dbd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:51Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.584191 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad406c24472c9ef7da82485a0600cf54f9002b83c0770c2ac86eca0f8779a870\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:51Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.595870 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"844146aa-36ac-4bee-8015-0d9850d94c4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cafb41d365e674a55638718df4e540ff77b6bd8e0d2d94bdb480b6617c0942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8f8a8bafbf88541375edb7c1b2134dec31a0d5cd46462a3271d24131bed4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hs9vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:51Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.634542 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.634587 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.634596 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.634611 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.634621 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:51Z","lastTransitionTime":"2025-10-11T03:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.736939 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.736965 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.736973 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.736985 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.737012 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:51Z","lastTransitionTime":"2025-10-11T03:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.771704 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d22c953c-0315-406e-9c5d-f1759837750e-metrics-certs\") pod \"network-metrics-daemon-p24x8\" (UID: \"d22c953c-0315-406e-9c5d-f1759837750e\") " pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:51:51 crc kubenswrapper[4967]: E1011 03:51:51.771934 4967 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 11 03:51:51 crc kubenswrapper[4967]: E1011 03:51:51.772059 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d22c953c-0315-406e-9c5d-f1759837750e-metrics-certs podName:d22c953c-0315-406e-9c5d-f1759837750e nodeName:}" failed. No retries permitted until 2025-10-11 03:51:55.77203494 +0000 UTC m=+43.735243903 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d22c953c-0315-406e-9c5d-f1759837750e-metrics-certs") pod "network-metrics-daemon-p24x8" (UID: "d22c953c-0315-406e-9c5d-f1759837750e") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.814022 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.814182 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.814208 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:51:51 crc kubenswrapper[4967]: E1011 03:51:51.814353 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.814397 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:51:51 crc kubenswrapper[4967]: E1011 03:51:51.814558 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:51:51 crc kubenswrapper[4967]: E1011 03:51:51.814792 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:51:51 crc kubenswrapper[4967]: E1011 03:51:51.814919 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.839787 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.839906 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.839925 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.839948 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.839965 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:51Z","lastTransitionTime":"2025-10-11T03:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.942993 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.943051 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.943067 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.943107 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:51 crc kubenswrapper[4967]: I1011 03:51:51.943121 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:51Z","lastTransitionTime":"2025-10-11T03:51:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.046267 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.046320 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.046341 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.046391 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.046415 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:52Z","lastTransitionTime":"2025-10-11T03:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.149359 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.149419 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.149436 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.149459 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.149476 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:52Z","lastTransitionTime":"2025-10-11T03:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.252997 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.253101 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.253120 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.253144 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.253162 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:52Z","lastTransitionTime":"2025-10-11T03:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.355258 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.355325 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.355342 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.355367 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.355384 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:52Z","lastTransitionTime":"2025-10-11T03:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.457772 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.457835 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.457853 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.457893 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.457909 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:52Z","lastTransitionTime":"2025-10-11T03:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.560906 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.560969 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.560987 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.561014 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.561038 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:52Z","lastTransitionTime":"2025-10-11T03:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.663368 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.663446 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.663471 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.663502 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.663524 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:52Z","lastTransitionTime":"2025-10-11T03:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.766766 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.766856 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.766880 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.766919 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.766947 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:52Z","lastTransitionTime":"2025-10-11T03:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.838407 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:52Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.858816 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:52Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.871234 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.871287 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.871305 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.871328 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.871345 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:52Z","lastTransitionTime":"2025-10-11T03:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.880064 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:52Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.898518 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:52Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.931182 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1c040b1dc41b86ac58bb6fe7d65c41fcac12bfcdf706e42a8cbf3389f90c09f3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c040b1dc41b86ac58bb6fe7d65c41fcac12bfcdf706e42a8cbf3389f90c09f3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:51:45Z\\\",\\\"message\\\":\\\"start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:45Z is after 2025-08-24T17:21:41Z]\\\\nI1011 03:51:45.040782 6396 services_controller.go:434] Service openshift-console-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-console-operator e4559ce3-2d5a-470f-b8bf-4c8b054d2335 11843 0 2025-02-23 05:38:55 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:console-operator] map[capability.openshift.io/name:Console include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc006efe87f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-lw6dx_openshift-ovn-kubernetes(270ac8da-b130-4f4f-bbf7-655229b64446)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:52Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.949058 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p24x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22c953c-0315-406e-9c5d-f1759837750e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p24x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:52Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.966784 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:52Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.973338 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.973389 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.973408 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.973431 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.973452 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:52Z","lastTransitionTime":"2025-10-11T03:51:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:52 crc kubenswrapper[4967]: I1011 03:51:52.982790 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:52Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.003460 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.033365 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.049370 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.069095 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.076531 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.076605 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.076632 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.076700 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.076728 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:53Z","lastTransitionTime":"2025-10-11T03:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.089595 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.111908 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.134117 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://257b2303215c18aa4500478fbfa39f9e1b95678c6d39ea51e4324c58f51dbd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.145576 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad406c24472c9ef7da82485a0600cf54f9002b83c0770c2ac86eca0f8779a870\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.160452 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"844146aa-36ac-4bee-8015-0d9850d94c4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cafb41d365e674a55638718df4e540ff77b6bd8e0d2d94bdb480b6617c0942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8f8a8bafbf88541375edb7c1b2134dec31a0d5cd46462a3271d24131bed4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hs9vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.178632 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.178686 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.178697 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.178713 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.178727 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:53Z","lastTransitionTime":"2025-10-11T03:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.282221 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.282330 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.282352 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.282419 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.282441 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:53Z","lastTransitionTime":"2025-10-11T03:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.385500 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.385555 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.385571 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.385595 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.385614 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:53Z","lastTransitionTime":"2025-10-11T03:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.489154 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.489240 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.489266 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.489296 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.489320 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:53Z","lastTransitionTime":"2025-10-11T03:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.592732 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.592793 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.592809 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.592833 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.592850 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:53Z","lastTransitionTime":"2025-10-11T03:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.695878 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.695937 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.695955 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.695982 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.695998 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:53Z","lastTransitionTime":"2025-10-11T03:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.799043 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.799134 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.799153 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.799179 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.799197 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:53Z","lastTransitionTime":"2025-10-11T03:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.814708 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.814785 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.814805 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.814708 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:51:53 crc kubenswrapper[4967]: E1011 03:51:53.814928 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:51:53 crc kubenswrapper[4967]: E1011 03:51:53.815064 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:51:53 crc kubenswrapper[4967]: E1011 03:51:53.815230 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:51:53 crc kubenswrapper[4967]: E1011 03:51:53.815384 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.902689 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.902750 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.902767 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.902791 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:53 crc kubenswrapper[4967]: I1011 03:51:53.902808 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:53Z","lastTransitionTime":"2025-10-11T03:51:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.005028 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.005170 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.005193 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.005224 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.005246 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:54Z","lastTransitionTime":"2025-10-11T03:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.108053 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.108098 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.108108 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.108120 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.108130 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:54Z","lastTransitionTime":"2025-10-11T03:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.211036 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.211137 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.211165 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.211193 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.211215 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:54Z","lastTransitionTime":"2025-10-11T03:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.314518 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.314595 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.314615 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.314640 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.314661 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:54Z","lastTransitionTime":"2025-10-11T03:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.417924 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.417996 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.418023 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.418052 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.418106 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:54Z","lastTransitionTime":"2025-10-11T03:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.521854 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.521914 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.521938 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.521971 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.521995 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:54Z","lastTransitionTime":"2025-10-11T03:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.625354 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.625416 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.625436 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.625495 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.625513 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:54Z","lastTransitionTime":"2025-10-11T03:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.728923 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.728986 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.729003 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.729029 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.729048 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:54Z","lastTransitionTime":"2025-10-11T03:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.832875 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.832947 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.832971 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.833007 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.833032 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:54Z","lastTransitionTime":"2025-10-11T03:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.936311 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.936381 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.936401 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.936428 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:54 crc kubenswrapper[4967]: I1011 03:51:54.936446 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:54Z","lastTransitionTime":"2025-10-11T03:51:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.039815 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.039866 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.039878 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.039896 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.039907 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:55Z","lastTransitionTime":"2025-10-11T03:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.143485 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.143575 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.143598 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.143625 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.143676 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:55Z","lastTransitionTime":"2025-10-11T03:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.246269 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.246313 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.246329 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.246352 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.246369 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:55Z","lastTransitionTime":"2025-10-11T03:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.349546 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.349612 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.349632 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.349656 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.349674 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:55Z","lastTransitionTime":"2025-10-11T03:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.452822 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.452880 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.452897 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.452922 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.452938 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:55Z","lastTransitionTime":"2025-10-11T03:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.556446 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.556505 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.556529 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.556561 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.556584 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:55Z","lastTransitionTime":"2025-10-11T03:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.660113 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.660168 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.660185 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.660209 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.660229 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:55Z","lastTransitionTime":"2025-10-11T03:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.763531 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.763598 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.763617 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.763642 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.763661 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:55Z","lastTransitionTime":"2025-10-11T03:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.814600 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.814674 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.814815 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:51:55 crc kubenswrapper[4967]: E1011 03:51:55.814806 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.814865 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:51:55 crc kubenswrapper[4967]: E1011 03:51:55.815056 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:51:55 crc kubenswrapper[4967]: E1011 03:51:55.815201 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:51:55 crc kubenswrapper[4967]: E1011 03:51:55.815378 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.818407 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d22c953c-0315-406e-9c5d-f1759837750e-metrics-certs\") pod \"network-metrics-daemon-p24x8\" (UID: \"d22c953c-0315-406e-9c5d-f1759837750e\") " pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:51:55 crc kubenswrapper[4967]: E1011 03:51:55.818663 4967 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 11 03:51:55 crc kubenswrapper[4967]: E1011 03:51:55.818786 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d22c953c-0315-406e-9c5d-f1759837750e-metrics-certs podName:d22c953c-0315-406e-9c5d-f1759837750e nodeName:}" failed. No retries permitted until 2025-10-11 03:52:03.818749035 +0000 UTC m=+51.781958018 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d22c953c-0315-406e-9c5d-f1759837750e-metrics-certs") pod "network-metrics-daemon-p24x8" (UID: "d22c953c-0315-406e-9c5d-f1759837750e") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.870023 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.870488 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.870572 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.870607 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.870665 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:55Z","lastTransitionTime":"2025-10-11T03:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.973647 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.973695 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.973714 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.973736 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:55 crc kubenswrapper[4967]: I1011 03:51:55.973753 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:55Z","lastTransitionTime":"2025-10-11T03:51:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.077426 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.077486 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.077506 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.077530 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.077548 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:56Z","lastTransitionTime":"2025-10-11T03:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.181217 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.181275 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.181291 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.181314 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.181333 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:56Z","lastTransitionTime":"2025-10-11T03:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.285027 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.285107 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.285126 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.285151 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.285168 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:56Z","lastTransitionTime":"2025-10-11T03:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.387871 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.387935 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.387955 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.387978 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.387995 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:56Z","lastTransitionTime":"2025-10-11T03:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.490857 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.490918 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.490939 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.490966 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.490984 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:56Z","lastTransitionTime":"2025-10-11T03:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.593638 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.593697 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.593715 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.593742 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.593764 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:56Z","lastTransitionTime":"2025-10-11T03:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.697048 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.697163 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.697188 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.697219 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.697239 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:56Z","lastTransitionTime":"2025-10-11T03:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.800485 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.800558 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.800582 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.800612 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.800636 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:56Z","lastTransitionTime":"2025-10-11T03:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.903477 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.903533 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.903553 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.903580 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:56 crc kubenswrapper[4967]: I1011 03:51:56.903598 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:56Z","lastTransitionTime":"2025-10-11T03:51:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.007009 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.007055 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.007066 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.007100 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.007114 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:57Z","lastTransitionTime":"2025-10-11T03:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.109715 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.109778 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.109794 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.109820 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.109837 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:57Z","lastTransitionTime":"2025-10-11T03:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.213797 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.213874 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.213892 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.213924 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.213944 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:57Z","lastTransitionTime":"2025-10-11T03:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.317705 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.317769 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.317814 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.317839 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.317858 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:57Z","lastTransitionTime":"2025-10-11T03:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.421443 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.421531 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.421550 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.421584 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.421607 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:57Z","lastTransitionTime":"2025-10-11T03:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.525183 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.525231 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.525248 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.525272 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.525286 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:57Z","lastTransitionTime":"2025-10-11T03:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.627733 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.627797 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.627807 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.627830 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.627844 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:57Z","lastTransitionTime":"2025-10-11T03:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.731419 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.731485 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.731506 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.731533 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.731548 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:57Z","lastTransitionTime":"2025-10-11T03:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.814619 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:51:57 crc kubenswrapper[4967]: E1011 03:51:57.814828 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.815595 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:51:57 crc kubenswrapper[4967]: E1011 03:51:57.815761 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.815838 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.815900 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:51:57 crc kubenswrapper[4967]: E1011 03:51:57.815981 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:51:57 crc kubenswrapper[4967]: E1011 03:51:57.816190 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.835065 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.835199 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.835220 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.835245 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.835262 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:57Z","lastTransitionTime":"2025-10-11T03:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.943028 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.943136 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.943160 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.943223 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:57 crc kubenswrapper[4967]: I1011 03:51:57.943245 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:57Z","lastTransitionTime":"2025-10-11T03:51:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.046842 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.046897 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.046914 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.046935 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.046953 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:58Z","lastTransitionTime":"2025-10-11T03:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.149763 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.149831 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.149850 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.149875 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.149897 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:58Z","lastTransitionTime":"2025-10-11T03:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.253293 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.253348 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.253359 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.253379 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.253398 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:58Z","lastTransitionTime":"2025-10-11T03:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.357277 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.357385 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.357452 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.357491 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.357567 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:58Z","lastTransitionTime":"2025-10-11T03:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.460340 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.460410 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.460433 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.460465 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.460491 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:58Z","lastTransitionTime":"2025-10-11T03:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.563767 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.563840 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.563863 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.563895 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.563917 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:58Z","lastTransitionTime":"2025-10-11T03:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.667067 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.667236 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.667266 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.667288 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.667304 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:58Z","lastTransitionTime":"2025-10-11T03:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.769667 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.769694 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.769702 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.769715 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.769724 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:58Z","lastTransitionTime":"2025-10-11T03:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.819012 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.819112 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.819129 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.819151 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.819165 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:58Z","lastTransitionTime":"2025-10-11T03:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:58 crc kubenswrapper[4967]: E1011 03:51:58.833152 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:58Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.838392 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.838457 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.838480 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.838503 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.838521 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:58Z","lastTransitionTime":"2025-10-11T03:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:58 crc kubenswrapper[4967]: E1011 03:51:58.853348 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:58Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.857663 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.857703 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.857712 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.857728 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.857739 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:58Z","lastTransitionTime":"2025-10-11T03:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:58 crc kubenswrapper[4967]: E1011 03:51:58.871574 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:58Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.875366 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.875418 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.875434 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.875457 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.875473 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:58Z","lastTransitionTime":"2025-10-11T03:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:58 crc kubenswrapper[4967]: E1011 03:51:58.887521 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:58Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.893095 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.893152 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.893162 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.893182 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.893193 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:58Z","lastTransitionTime":"2025-10-11T03:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:58 crc kubenswrapper[4967]: E1011 03:51:58.906732 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:58Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:58Z is after 2025-08-24T17:21:41Z" Oct 11 03:51:58 crc kubenswrapper[4967]: E1011 03:51:58.906889 4967 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.908527 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.908566 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.908576 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.908593 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:58 crc kubenswrapper[4967]: I1011 03:51:58.908603 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:58Z","lastTransitionTime":"2025-10-11T03:51:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.011032 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.011088 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.011097 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.011110 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.011122 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:59Z","lastTransitionTime":"2025-10-11T03:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.113836 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.113898 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.113915 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.113938 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.113956 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:59Z","lastTransitionTime":"2025-10-11T03:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.217483 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.217546 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.217566 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.217590 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.217610 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:59Z","lastTransitionTime":"2025-10-11T03:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.320716 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.320808 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.320836 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.320864 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.320887 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:59Z","lastTransitionTime":"2025-10-11T03:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.423787 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.423861 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.423878 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.423903 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.423921 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:59Z","lastTransitionTime":"2025-10-11T03:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.526534 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.526585 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.526604 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.526627 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.526643 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:59Z","lastTransitionTime":"2025-10-11T03:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.629196 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.629273 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.629289 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.629315 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.629333 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:59Z","lastTransitionTime":"2025-10-11T03:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.732852 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.732909 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.732925 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.732948 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.732964 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:59Z","lastTransitionTime":"2025-10-11T03:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.814645 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.815245 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.815271 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:51:59 crc kubenswrapper[4967]: E1011 03:51:59.815551 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.815762 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:51:59 crc kubenswrapper[4967]: E1011 03:51:59.816105 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:51:59 crc kubenswrapper[4967]: E1011 03:51:59.816332 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:51:59 crc kubenswrapper[4967]: E1011 03:51:59.816601 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.818541 4967 scope.go:117] "RemoveContainer" containerID="1c040b1dc41b86ac58bb6fe7d65c41fcac12bfcdf706e42a8cbf3389f90c09f3" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.835228 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.835274 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.835294 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.835317 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.835335 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:59Z","lastTransitionTime":"2025-10-11T03:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.938403 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.938448 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.938464 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.938487 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:51:59 crc kubenswrapper[4967]: I1011 03:51:59.938503 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:51:59Z","lastTransitionTime":"2025-10-11T03:51:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.041957 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.042374 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.042392 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.042417 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.042438 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:00Z","lastTransitionTime":"2025-10-11T03:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.141315 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lw6dx_270ac8da-b130-4f4f-bbf7-655229b64446/ovnkube-controller/1.log" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.144257 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.144298 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.144315 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.144361 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.144377 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:00Z","lastTransitionTime":"2025-10-11T03:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.146359 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" event={"ID":"270ac8da-b130-4f4f-bbf7-655229b64446","Type":"ContainerStarted","Data":"3adee99370856f7d1dd95c3fed98740090f99e3d25c449f6f36170f0aeacd3e9"} Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.147110 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.168628 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:00Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.180793 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:00Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.197762 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:00Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.212884 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:00Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.234768 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3adee99370856f7d1dd95c3fed98740090f99e3d25c449f6f36170f0aeacd3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c040b1dc41b86ac58bb6fe7d65c41fcac12bfcdf706e42a8cbf3389f90c09f3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:51:45Z\\\",\\\"message\\\":\\\"start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:45Z is after 2025-08-24T17:21:41Z]\\\\nI1011 03:51:45.040782 6396 services_controller.go:434] Service openshift-console-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-console-operator e4559ce3-2d5a-470f-b8bf-4c8b054d2335 11843 0 2025-02-23 05:38:55 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:console-operator] map[capability.openshift.io/name:Console include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc006efe87f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:00Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.246803 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.246851 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.246861 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.246877 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.246889 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:00Z","lastTransitionTime":"2025-10-11T03:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.248569 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p24x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22c953c-0315-406e-9c5d-f1759837750e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p24x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:00Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.270008 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:00Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.284579 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:00Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.298116 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:00Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.327132 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:00Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.343048 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:00Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.348889 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.348944 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.348962 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.348985 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.349001 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:00Z","lastTransitionTime":"2025-10-11T03:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.365490 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:00Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.385061 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:00Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.399962 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:00Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.420707 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://257b2303215c18aa4500478fbfa39f9e1b95678c6d39ea51e4324c58f51dbd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:00Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.431425 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad406c24472c9ef7da82485a0600cf54f9002b83c0770c2ac86eca0f8779a870\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:00Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.446604 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"844146aa-36ac-4bee-8015-0d9850d94c4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cafb41d365e674a55638718df4e540ff77b6bd8e0d2d94bdb480b6617c0942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8f8a8bafbf88541375edb7c1b2134dec31a0d5cd46462a3271d24131bed4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hs9vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:00Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.451307 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.451376 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.451401 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.451434 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.451459 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:00Z","lastTransitionTime":"2025-10-11T03:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.553805 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.554900 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.554990 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.555259 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.555346 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:00Z","lastTransitionTime":"2025-10-11T03:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.658097 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.658146 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.658163 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.658185 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.658199 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:00Z","lastTransitionTime":"2025-10-11T03:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.761550 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.761621 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.761639 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.761669 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.761691 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:00Z","lastTransitionTime":"2025-10-11T03:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.864036 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.864098 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.864110 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.864127 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.864139 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:00Z","lastTransitionTime":"2025-10-11T03:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.967031 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.967128 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.967149 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.967174 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:00 crc kubenswrapper[4967]: I1011 03:52:00.967194 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:00Z","lastTransitionTime":"2025-10-11T03:52:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.070442 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.070504 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.070522 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.070547 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.070565 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:01Z","lastTransitionTime":"2025-10-11T03:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.153014 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lw6dx_270ac8da-b130-4f4f-bbf7-655229b64446/ovnkube-controller/2.log" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.154114 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lw6dx_270ac8da-b130-4f4f-bbf7-655229b64446/ovnkube-controller/1.log" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.158859 4967 generic.go:334] "Generic (PLEG): container finished" podID="270ac8da-b130-4f4f-bbf7-655229b64446" containerID="3adee99370856f7d1dd95c3fed98740090f99e3d25c449f6f36170f0aeacd3e9" exitCode=1 Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.158933 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" event={"ID":"270ac8da-b130-4f4f-bbf7-655229b64446","Type":"ContainerDied","Data":"3adee99370856f7d1dd95c3fed98740090f99e3d25c449f6f36170f0aeacd3e9"} Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.158997 4967 scope.go:117] "RemoveContainer" containerID="1c040b1dc41b86ac58bb6fe7d65c41fcac12bfcdf706e42a8cbf3389f90c09f3" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.160314 4967 scope.go:117] "RemoveContainer" containerID="3adee99370856f7d1dd95c3fed98740090f99e3d25c449f6f36170f0aeacd3e9" Oct 11 03:52:01 crc kubenswrapper[4967]: E1011 03:52:01.160659 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lw6dx_openshift-ovn-kubernetes(270ac8da-b130-4f4f-bbf7-655229b64446)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.174417 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.174491 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.174514 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.174545 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.174568 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:01Z","lastTransitionTime":"2025-10-11T03:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.196431 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3adee99370856f7d1dd95c3fed98740090f99e3d25c449f6f36170f0aeacd3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1c040b1dc41b86ac58bb6fe7d65c41fcac12bfcdf706e42a8cbf3389f90c09f3\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:51:45Z\\\",\\\"message\\\":\\\"start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:51:45Z is after 2025-08-24T17:21:41Z]\\\\nI1011 03:51:45.040782 6396 services_controller.go:434] Service openshift-console-operator/metrics retrieved from lister for network=default: \\\\u0026Service{ObjectMeta:{metrics openshift-console-operator e4559ce3-2d5a-470f-b8bf-4c8b054d2335 11843 0 2025-02-23 05:38:55 +0000 UTC \\\\u003cnil\\\\u003e \\\\u003cnil\\\\u003e map[name:console-operator] map[capability.openshift.io/name:Console include.release.openshift.io/hypershift:true include.release.openshift.io/ibm-cloud-managed:true include.release.openshift.io/self-managed-high-availability:true include.release.openshift.io/single-node-developer:true service.alpha.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168 service.beta.openshift.io/serving-cert-secret-name:serving-cert service.beta.openshift.io/serving-cert-signed-by:openshift-service-serving-signer@1740288168] [{config.openshift.io/v1 ClusterVersion version 9101b518-476b-4eea-8fa6-69b0534e5caa 0xc006efe87f \\\\u003cnil\\\\u003e}] [] []},Spec:ServiceSpec{Ports:[]ServicePort\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:44Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3adee99370856f7d1dd95c3fed98740090f99e3d25c449f6f36170f0aeacd3e9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:52:00Z\\\",\\\"message\\\":\\\":311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:52:00.856605 6612 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.856659 6612 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.856768 6612 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.857120 6612 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:52:00.857719 6612 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1011 03:52:00.857743 6612 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1011 03:52:00.857756 6612 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1011 03:52:00.857789 6612 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1011 03:52:00.857813 6612 factory.go:656] Stopping watch factory\\\\nI1011 03:52:00.857815 6612 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1011 03:52:00.857835 6612 ovnkube.go:599] Stopped ovnkube\\\\nI1011 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:59Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:01Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.210722 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p24x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22c953c-0315-406e-9c5d-f1759837750e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p24x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:01Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.224359 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:01Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.244345 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:01Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.261862 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:01Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.277272 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.277637 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.277865 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.278113 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.278359 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:01Z","lastTransitionTime":"2025-10-11T03:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.280198 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:01Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.297928 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:01Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.311281 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:01Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.331578 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:01Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.356460 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:01Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.373817 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad406c24472c9ef7da82485a0600cf54f9002b83c0770c2ac86eca0f8779a870\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:01Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.381418 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.381468 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.381482 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.381502 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.381519 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:01Z","lastTransitionTime":"2025-10-11T03:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.390843 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"844146aa-36ac-4bee-8015-0d9850d94c4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cafb41d365e674a55638718df4e540ff77b6bd8e0d2d94bdb480b6617c0942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8f8a8bafbf88541375edb7c1b2134dec31a0d5cd46462a3271d24131bed4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hs9vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:01Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.413829 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://257b2303215c18aa4500478fbfa39f9e1b95678c6d39ea51e4324c58f51dbd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:01Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.433592 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:01Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.452959 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:01Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.469350 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:01Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.483928 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.484001 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.484020 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.484103 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.484125 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:01Z","lastTransitionTime":"2025-10-11T03:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.488885 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:01Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.585867 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.585923 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.585936 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.585955 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.585969 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:01Z","lastTransitionTime":"2025-10-11T03:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.689619 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.689694 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.689715 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.689745 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.689768 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:01Z","lastTransitionTime":"2025-10-11T03:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.792383 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.792466 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.792488 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.792522 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.792545 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:01Z","lastTransitionTime":"2025-10-11T03:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.814687 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.814776 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:52:01 crc kubenswrapper[4967]: E1011 03:52:01.814794 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.814881 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.814910 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:52:01 crc kubenswrapper[4967]: E1011 03:52:01.815126 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:52:01 crc kubenswrapper[4967]: E1011 03:52:01.815207 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:52:01 crc kubenswrapper[4967]: E1011 03:52:01.815343 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.895355 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.895393 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.895401 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.895415 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.895424 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:01Z","lastTransitionTime":"2025-10-11T03:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.997930 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.998005 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.998028 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.998056 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:01 crc kubenswrapper[4967]: I1011 03:52:01.998111 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:01Z","lastTransitionTime":"2025-10-11T03:52:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.100815 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.100864 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.100875 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.100891 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.100903 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:02Z","lastTransitionTime":"2025-10-11T03:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.164197 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lw6dx_270ac8da-b130-4f4f-bbf7-655229b64446/ovnkube-controller/2.log" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.168771 4967 scope.go:117] "RemoveContainer" containerID="3adee99370856f7d1dd95c3fed98740090f99e3d25c449f6f36170f0aeacd3e9" Oct 11 03:52:02 crc kubenswrapper[4967]: E1011 03:52:02.169158 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lw6dx_openshift-ovn-kubernetes(270ac8da-b130-4f4f-bbf7-655229b64446)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.189191 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3adee99370856f7d1dd95c3fed98740090f99e3d25c449f6f36170f0aeacd3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3adee99370856f7d1dd95c3fed98740090f99e3d25c449f6f36170f0aeacd3e9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:52:00Z\\\",\\\"message\\\":\\\":311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:52:00.856605 6612 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.856659 6612 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.856768 6612 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.857120 6612 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:52:00.857719 6612 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1011 03:52:00.857743 6612 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1011 03:52:00.857756 6612 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1011 03:52:00.857789 6612 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1011 03:52:00.857813 6612 factory.go:656] Stopping watch factory\\\\nI1011 03:52:00.857815 6612 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1011 03:52:00.857835 6612 ovnkube.go:599] Stopped ovnkube\\\\nI1011 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lw6dx_openshift-ovn-kubernetes(270ac8da-b130-4f4f-bbf7-655229b64446)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.203230 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.203292 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.203311 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.203339 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.203359 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:02Z","lastTransitionTime":"2025-10-11T03:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.206551 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p24x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22c953c-0315-406e-9c5d-f1759837750e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p24x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.219688 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.252165 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.269736 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.286208 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.301351 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.306304 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.306336 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.306347 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.306365 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.306380 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:02Z","lastTransitionTime":"2025-10-11T03:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.316282 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.327157 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.337098 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.353975 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://257b2303215c18aa4500478fbfa39f9e1b95678c6d39ea51e4324c58f51dbd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.363903 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad406c24472c9ef7da82485a0600cf54f9002b83c0770c2ac86eca0f8779a870\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.379156 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"844146aa-36ac-4bee-8015-0d9850d94c4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cafb41d365e674a55638718df4e540ff77b6bd8e0d2d94bdb480b6617c0942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8f8a8bafbf88541375edb7c1b2134dec31a0d5cd46462a3271d24131bed4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hs9vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.399573 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.408925 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.408961 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.408970 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.408985 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.408996 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:02Z","lastTransitionTime":"2025-10-11T03:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.415588 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.427665 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.441137 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.510564 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.510610 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.510623 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.510644 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.510657 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:02Z","lastTransitionTime":"2025-10-11T03:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.612919 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.612957 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.612967 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.612983 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.612994 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:02Z","lastTransitionTime":"2025-10-11T03:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.623085 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.634665 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.647410 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3adee99370856f7d1dd95c3fed98740090f99e3d25c449f6f36170f0aeacd3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3adee99370856f7d1dd95c3fed98740090f99e3d25c449f6f36170f0aeacd3e9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:52:00Z\\\",\\\"message\\\":\\\":311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:52:00.856605 6612 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.856659 6612 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.856768 6612 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.857120 6612 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:52:00.857719 6612 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1011 03:52:00.857743 6612 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1011 03:52:00.857756 6612 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1011 03:52:00.857789 6612 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1011 03:52:00.857813 6612 factory.go:656] Stopping watch factory\\\\nI1011 03:52:00.857815 6612 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1011 03:52:00.857835 6612 ovnkube.go:599] Stopped ovnkube\\\\nI1011 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lw6dx_openshift-ovn-kubernetes(270ac8da-b130-4f4f-bbf7-655229b64446)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.658428 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p24x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22c953c-0315-406e-9c5d-f1759837750e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p24x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.674819 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.689841 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.702389 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.715790 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.716201 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.716416 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.716623 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.716115 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.716833 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:02Z","lastTransitionTime":"2025-10-11T03:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.739421 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.759983 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.771772 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.783099 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.800472 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://257b2303215c18aa4500478fbfa39f9e1b95678c6d39ea51e4324c58f51dbd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.811391 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad406c24472c9ef7da82485a0600cf54f9002b83c0770c2ac86eca0f8779a870\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.819225 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.819254 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.819262 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.819276 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.819285 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:02Z","lastTransitionTime":"2025-10-11T03:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.825675 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"844146aa-36ac-4bee-8015-0d9850d94c4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cafb41d365e674a55638718df4e540ff77b6bd8e0d2d94bdb480b6617c0942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8f8a8bafbf88541375edb7c1b2134dec31a0d5cd46462a3271d24131bed4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hs9vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.839929 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.854154 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.872908 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.889975 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.904213 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b09975d-4c55-4b1a-b034-851ebf64cd3f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b90baa6dbb24ee4c2deea587bada1820917fd963caaf91210c25409b254c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fef6a4dbe93461c6352741acfc9439a22668f46dd725fcf80fb37d93c88e9d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f23f4ffaeac5c8606def100262484a45cd86296beb91735215c57955e45067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df874a11e199cea6058abe0e5bb63faf7246199fa02093f05720bc41dbb7a81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df874a11e199cea6058abe0e5bb63faf7246199fa02093f05720bc41dbb7a81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.921846 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.921923 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.921941 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.921968 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.921985 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:02Z","lastTransitionTime":"2025-10-11T03:52:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.926173 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3adee99370856f7d1dd95c3fed98740090f99e3d25c449f6f36170f0aeacd3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3adee99370856f7d1dd95c3fed98740090f99e3d25c449f6f36170f0aeacd3e9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:52:00Z\\\",\\\"message\\\":\\\":311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:52:00.856605 6612 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.856659 6612 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.856768 6612 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.857120 6612 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:52:00.857719 6612 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1011 03:52:00.857743 6612 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1011 03:52:00.857756 6612 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1011 03:52:00.857789 6612 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1011 03:52:00.857813 6612 factory.go:656] Stopping watch factory\\\\nI1011 03:52:00.857815 6612 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1011 03:52:00.857835 6612 ovnkube.go:599] Stopped ovnkube\\\\nI1011 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lw6dx_openshift-ovn-kubernetes(270ac8da-b130-4f4f-bbf7-655229b64446)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.942652 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p24x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22c953c-0315-406e-9c5d-f1759837750e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p24x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.956359 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:02 crc kubenswrapper[4967]: I1011 03:52:02.988268 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:02Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.002888 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:03Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.022247 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:03Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.024089 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.024231 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.024345 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.024438 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.024549 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:03Z","lastTransitionTime":"2025-10-11T03:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.037955 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:03Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.059045 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:03Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.077255 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:03Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.091954 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:03Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.112585 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://257b2303215c18aa4500478fbfa39f9e1b95678c6d39ea51e4324c58f51dbd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:03Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.125187 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad406c24472c9ef7da82485a0600cf54f9002b83c0770c2ac86eca0f8779a870\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:03Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.128649 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.128710 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.128729 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.128759 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.128777 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:03Z","lastTransitionTime":"2025-10-11T03:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.142787 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"844146aa-36ac-4bee-8015-0d9850d94c4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cafb41d365e674a55638718df4e540ff77b6bd8e0d2d94bdb480b6617c0942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8f8a8bafbf88541375edb7c1b2134dec31a0d5cd46462a3271d24131bed4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hs9vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:03Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.164688 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:03Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.185000 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:03Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.198056 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:03Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.209332 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:03Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.231121 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.231178 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.231196 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.231223 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.231241 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:03Z","lastTransitionTime":"2025-10-11T03:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.334431 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.334495 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.334517 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.334547 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.334570 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:03Z","lastTransitionTime":"2025-10-11T03:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.437896 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.438498 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.438734 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.438951 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.439184 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:03Z","lastTransitionTime":"2025-10-11T03:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.542274 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.542345 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.542363 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.542386 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.542405 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:03Z","lastTransitionTime":"2025-10-11T03:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.611068 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.611192 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:03 crc kubenswrapper[4967]: E1011 03:52:03.611375 4967 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 11 03:52:03 crc kubenswrapper[4967]: E1011 03:52:03.611526 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-11 03:52:35.611495029 +0000 UTC m=+83.574704002 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 11 03:52:03 crc kubenswrapper[4967]: E1011 03:52:03.611395 4967 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 11 03:52:03 crc kubenswrapper[4967]: E1011 03:52:03.611610 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-11 03:52:35.611596281 +0000 UTC m=+83.574805254 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.644934 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.644978 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.644986 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.645001 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.645010 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:03Z","lastTransitionTime":"2025-10-11T03:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.712502 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:52:03 crc kubenswrapper[4967]: E1011 03:52:03.712807 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 11 03:52:03 crc kubenswrapper[4967]: E1011 03:52:03.713125 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 11 03:52:03 crc kubenswrapper[4967]: E1011 03:52:03.713156 4967 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:52:03 crc kubenswrapper[4967]: E1011 03:52:03.713249 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-11 03:52:35.713219981 +0000 UTC m=+83.676428954 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.713011 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:52:03 crc kubenswrapper[4967]: E1011 03:52:03.713798 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 11 03:52:03 crc kubenswrapper[4967]: E1011 03:52:03.714007 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 11 03:52:03 crc kubenswrapper[4967]: E1011 03:52:03.714252 4967 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:52:03 crc kubenswrapper[4967]: E1011 03:52:03.714546 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-11 03:52:35.714520839 +0000 UTC m=+83.677729812 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.748588 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.748636 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.748655 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.748678 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.748696 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:03Z","lastTransitionTime":"2025-10-11T03:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.814269 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:52:03 crc kubenswrapper[4967]: E1011 03:52:03.814626 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:52:35.814600267 +0000 UTC m=+83.777809240 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.814579 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.814616 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.814527 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.814713 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:03 crc kubenswrapper[4967]: E1011 03:52:03.814907 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:52:03 crc kubenswrapper[4967]: E1011 03:52:03.815023 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:52:03 crc kubenswrapper[4967]: E1011 03:52:03.815216 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:52:03 crc kubenswrapper[4967]: E1011 03:52:03.815336 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.851786 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.851851 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.851874 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.851904 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.851925 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:03Z","lastTransitionTime":"2025-10-11T03:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.915716 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d22c953c-0315-406e-9c5d-f1759837750e-metrics-certs\") pod \"network-metrics-daemon-p24x8\" (UID: \"d22c953c-0315-406e-9c5d-f1759837750e\") " pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:03 crc kubenswrapper[4967]: E1011 03:52:03.915929 4967 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 11 03:52:03 crc kubenswrapper[4967]: E1011 03:52:03.916293 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d22c953c-0315-406e-9c5d-f1759837750e-metrics-certs podName:d22c953c-0315-406e-9c5d-f1759837750e nodeName:}" failed. No retries permitted until 2025-10-11 03:52:19.916261857 +0000 UTC m=+67.879470820 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d22c953c-0315-406e-9c5d-f1759837750e-metrics-certs") pod "network-metrics-daemon-p24x8" (UID: "d22c953c-0315-406e-9c5d-f1759837750e") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.955294 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.955377 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.955403 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.955437 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:03 crc kubenswrapper[4967]: I1011 03:52:03.955456 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:03Z","lastTransitionTime":"2025-10-11T03:52:03Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.058708 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.058762 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.058782 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.058805 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.058821 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:04Z","lastTransitionTime":"2025-10-11T03:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.161685 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.161738 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.161754 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.161777 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.161794 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:04Z","lastTransitionTime":"2025-10-11T03:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.265244 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.265303 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.265323 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.265346 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.265363 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:04Z","lastTransitionTime":"2025-10-11T03:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.369374 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.369448 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.369467 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.369496 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.369520 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:04Z","lastTransitionTime":"2025-10-11T03:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.473148 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.473220 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.473249 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.473277 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.473294 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:04Z","lastTransitionTime":"2025-10-11T03:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.575971 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.576034 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.576051 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.576118 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.576145 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:04Z","lastTransitionTime":"2025-10-11T03:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.678788 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.678865 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.678880 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.678902 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.678919 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:04Z","lastTransitionTime":"2025-10-11T03:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.782737 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.782926 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.782946 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.783020 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.783042 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:04Z","lastTransitionTime":"2025-10-11T03:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.886530 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.886780 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.886960 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.887099 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.887222 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:04Z","lastTransitionTime":"2025-10-11T03:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.990088 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.990124 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.990132 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.990145 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:04 crc kubenswrapper[4967]: I1011 03:52:04.990155 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:04Z","lastTransitionTime":"2025-10-11T03:52:04Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.092784 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.092831 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.092845 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.092864 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.092881 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:05Z","lastTransitionTime":"2025-10-11T03:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.196549 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.196921 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.197101 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.197264 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.197431 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:05Z","lastTransitionTime":"2025-10-11T03:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.302130 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.302216 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.302243 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.302278 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.302314 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:05Z","lastTransitionTime":"2025-10-11T03:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.404943 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.405017 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.405049 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.405109 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.405133 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:05Z","lastTransitionTime":"2025-10-11T03:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.507907 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.507974 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.507993 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.508017 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.508035 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:05Z","lastTransitionTime":"2025-10-11T03:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.610559 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.610621 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.610638 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.610664 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.610682 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:05Z","lastTransitionTime":"2025-10-11T03:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.713410 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.713475 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.713494 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.713520 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.713539 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:05Z","lastTransitionTime":"2025-10-11T03:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.814015 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.814041 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.814135 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.814170 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:52:05 crc kubenswrapper[4967]: E1011 03:52:05.814304 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:52:05 crc kubenswrapper[4967]: E1011 03:52:05.814503 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:52:05 crc kubenswrapper[4967]: E1011 03:52:05.814638 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:52:05 crc kubenswrapper[4967]: E1011 03:52:05.814741 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.816138 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.816182 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.816199 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.816223 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.816241 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:05Z","lastTransitionTime":"2025-10-11T03:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.919151 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.919219 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.919242 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.919270 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:05 crc kubenswrapper[4967]: I1011 03:52:05.919293 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:05Z","lastTransitionTime":"2025-10-11T03:52:05Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.022198 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.022240 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.022249 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.022262 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.022271 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:06Z","lastTransitionTime":"2025-10-11T03:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.125623 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.125659 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.125668 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.125681 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.125690 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:06Z","lastTransitionTime":"2025-10-11T03:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.228842 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.228899 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.228916 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.228941 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.228958 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:06Z","lastTransitionTime":"2025-10-11T03:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.330784 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.330845 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.330865 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.330889 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.330906 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:06Z","lastTransitionTime":"2025-10-11T03:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.433180 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.433244 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.433263 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.433292 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.433309 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:06Z","lastTransitionTime":"2025-10-11T03:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.536141 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.536196 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.536209 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.536229 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.536241 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:06Z","lastTransitionTime":"2025-10-11T03:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.639066 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.639168 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.639190 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.639219 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.639240 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:06Z","lastTransitionTime":"2025-10-11T03:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.741556 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.741613 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.741631 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.741657 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.741675 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:06Z","lastTransitionTime":"2025-10-11T03:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.845231 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.845677 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.845897 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.846131 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.846326 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:06Z","lastTransitionTime":"2025-10-11T03:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.949266 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.949345 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.949363 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.949402 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:06 crc kubenswrapper[4967]: I1011 03:52:06.949421 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:06Z","lastTransitionTime":"2025-10-11T03:52:06Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.052246 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.052315 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.052334 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.052358 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.052373 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:07Z","lastTransitionTime":"2025-10-11T03:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.156207 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.156272 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.156292 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.156322 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.156343 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:07Z","lastTransitionTime":"2025-10-11T03:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.259608 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.259694 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.259715 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.259743 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.259764 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:07Z","lastTransitionTime":"2025-10-11T03:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.362597 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.362664 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.362682 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.362709 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.362732 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:07Z","lastTransitionTime":"2025-10-11T03:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.465608 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.465678 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.465696 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.465722 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.465740 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:07Z","lastTransitionTime":"2025-10-11T03:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.569236 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.569336 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.569357 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.569414 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.569434 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:07Z","lastTransitionTime":"2025-10-11T03:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.677490 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.677572 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.677591 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.677615 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.677634 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:07Z","lastTransitionTime":"2025-10-11T03:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.781214 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.781284 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.781308 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.781338 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.781358 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:07Z","lastTransitionTime":"2025-10-11T03:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.814165 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.814202 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.814202 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.814205 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:52:07 crc kubenswrapper[4967]: E1011 03:52:07.814320 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:52:07 crc kubenswrapper[4967]: E1011 03:52:07.814591 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:52:07 crc kubenswrapper[4967]: E1011 03:52:07.814706 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:52:07 crc kubenswrapper[4967]: E1011 03:52:07.814821 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.883952 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.884027 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.884051 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.884129 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.884155 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:07Z","lastTransitionTime":"2025-10-11T03:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.986952 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.987038 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.987101 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.987133 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:07 crc kubenswrapper[4967]: I1011 03:52:07.987154 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:07Z","lastTransitionTime":"2025-10-11T03:52:07Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.090623 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.090724 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.090742 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.090770 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.090790 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:08Z","lastTransitionTime":"2025-10-11T03:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.192581 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.192630 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.192648 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.192669 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.192687 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:08Z","lastTransitionTime":"2025-10-11T03:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.296496 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.296573 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.296634 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.296664 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.296686 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:08Z","lastTransitionTime":"2025-10-11T03:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.400113 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.400177 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.400198 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.400227 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.400243 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:08Z","lastTransitionTime":"2025-10-11T03:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.502710 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.502768 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.502789 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.502817 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.502839 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:08Z","lastTransitionTime":"2025-10-11T03:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.605715 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.606022 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.606236 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.606394 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.606543 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:08Z","lastTransitionTime":"2025-10-11T03:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.709579 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.709711 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.709731 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.709755 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.709772 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:08Z","lastTransitionTime":"2025-10-11T03:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.812741 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.812781 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.812792 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.812809 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.812819 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:08Z","lastTransitionTime":"2025-10-11T03:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.915685 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.915756 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.915767 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.915786 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:08 crc kubenswrapper[4967]: I1011 03:52:08.915797 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:08Z","lastTransitionTime":"2025-10-11T03:52:08Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.018913 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.018986 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.019008 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.019036 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.019057 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:09Z","lastTransitionTime":"2025-10-11T03:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.093519 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.093585 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.093607 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.093638 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.093662 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:09Z","lastTransitionTime":"2025-10-11T03:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:09 crc kubenswrapper[4967]: E1011 03:52:09.116263 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:09Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.121068 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.121401 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.121655 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.121867 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.122065 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:09Z","lastTransitionTime":"2025-10-11T03:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:09 crc kubenswrapper[4967]: E1011 03:52:09.145842 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:09Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.151615 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.151969 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.152198 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.152407 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.152632 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:09Z","lastTransitionTime":"2025-10-11T03:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:09 crc kubenswrapper[4967]: E1011 03:52:09.174183 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:09Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.178137 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.178322 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.178466 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.178600 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.178725 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:09Z","lastTransitionTime":"2025-10-11T03:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:09 crc kubenswrapper[4967]: E1011 03:52:09.195821 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:09Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.205686 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.205735 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.205748 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.205765 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.205779 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:09Z","lastTransitionTime":"2025-10-11T03:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:09 crc kubenswrapper[4967]: E1011 03:52:09.223505 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:09Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:09Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:09 crc kubenswrapper[4967]: E1011 03:52:09.223641 4967 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.225478 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.225572 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.225599 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.225634 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.225660 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:09Z","lastTransitionTime":"2025-10-11T03:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.327731 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.327792 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.327813 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.327838 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.327857 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:09Z","lastTransitionTime":"2025-10-11T03:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.430558 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.430618 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.430639 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.430663 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.430680 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:09Z","lastTransitionTime":"2025-10-11T03:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.534228 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.534292 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.534320 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.534352 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.534373 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:09Z","lastTransitionTime":"2025-10-11T03:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.637023 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.637135 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.637155 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.637185 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.637204 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:09Z","lastTransitionTime":"2025-10-11T03:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.740208 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.740270 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.740291 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.740314 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.740331 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:09Z","lastTransitionTime":"2025-10-11T03:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.813817 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.813833 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:09 crc kubenswrapper[4967]: E1011 03:52:09.814001 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.813829 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.813829 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:52:09 crc kubenswrapper[4967]: E1011 03:52:09.814130 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:52:09 crc kubenswrapper[4967]: E1011 03:52:09.814247 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:52:09 crc kubenswrapper[4967]: E1011 03:52:09.814380 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.843144 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.843212 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.843231 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.843256 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.843276 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:09Z","lastTransitionTime":"2025-10-11T03:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.946710 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.946788 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.946813 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.946842 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:09 crc kubenswrapper[4967]: I1011 03:52:09.946862 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:09Z","lastTransitionTime":"2025-10-11T03:52:09Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.049783 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.049858 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.049882 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.049911 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.049933 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:10Z","lastTransitionTime":"2025-10-11T03:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.153321 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.153384 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.153406 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.153432 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.153453 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:10Z","lastTransitionTime":"2025-10-11T03:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.256976 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.257049 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.257066 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.257119 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.257139 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:10Z","lastTransitionTime":"2025-10-11T03:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.359762 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.359803 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.359820 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.359841 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.359858 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:10Z","lastTransitionTime":"2025-10-11T03:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.463176 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.463236 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.463251 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.463271 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.463286 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:10Z","lastTransitionTime":"2025-10-11T03:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.566240 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.566320 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.566339 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.566368 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.566387 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:10Z","lastTransitionTime":"2025-10-11T03:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.669024 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.669136 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.669160 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.669185 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.669203 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:10Z","lastTransitionTime":"2025-10-11T03:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.773211 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.773290 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.773305 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.773326 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.773342 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:10Z","lastTransitionTime":"2025-10-11T03:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.876616 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.876717 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.876739 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.876775 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.876792 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:10Z","lastTransitionTime":"2025-10-11T03:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.980305 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.980376 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.980400 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.980429 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:10 crc kubenswrapper[4967]: I1011 03:52:10.980454 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:10Z","lastTransitionTime":"2025-10-11T03:52:10Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.083478 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.083556 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.083577 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.083601 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.083620 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:11Z","lastTransitionTime":"2025-10-11T03:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.186435 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.186498 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.186515 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.186540 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.186564 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:11Z","lastTransitionTime":"2025-10-11T03:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.294345 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.294546 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.294775 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.294830 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.294869 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:11Z","lastTransitionTime":"2025-10-11T03:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.399528 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.399596 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.399617 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.399648 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.399671 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:11Z","lastTransitionTime":"2025-10-11T03:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.503293 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.503339 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.503350 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.503369 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.503384 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:11Z","lastTransitionTime":"2025-10-11T03:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.607706 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.607762 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.607778 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.607803 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.607819 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:11Z","lastTransitionTime":"2025-10-11T03:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.710269 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.710338 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.710394 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.710424 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.710448 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:11Z","lastTransitionTime":"2025-10-11T03:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.816969 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.817018 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.817043 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.817103 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.817058 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:11 crc kubenswrapper[4967]: E1011 03:52:11.817213 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.817231 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.817292 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.817320 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:11Z","lastTransitionTime":"2025-10-11T03:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.817418 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:52:11 crc kubenswrapper[4967]: E1011 03:52:11.817451 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:52:11 crc kubenswrapper[4967]: E1011 03:52:11.817611 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:52:11 crc kubenswrapper[4967]: E1011 03:52:11.817701 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.923488 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.923540 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.923560 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.923582 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:11 crc kubenswrapper[4967]: I1011 03:52:11.923599 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:11Z","lastTransitionTime":"2025-10-11T03:52:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.026873 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.026948 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.026972 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.027002 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.027023 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:12Z","lastTransitionTime":"2025-10-11T03:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.130840 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.130905 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.130923 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.130948 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.130964 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:12Z","lastTransitionTime":"2025-10-11T03:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.233393 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.233454 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.233478 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.233508 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.233529 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:12Z","lastTransitionTime":"2025-10-11T03:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.336430 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.336486 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.336504 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.336525 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.336544 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:12Z","lastTransitionTime":"2025-10-11T03:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.439581 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.439658 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.439683 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.439712 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.439730 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:12Z","lastTransitionTime":"2025-10-11T03:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.542653 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.542710 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.542727 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.542751 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.542769 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:12Z","lastTransitionTime":"2025-10-11T03:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.646181 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.646260 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.646279 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.646305 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.646324 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:12Z","lastTransitionTime":"2025-10-11T03:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.748913 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.748980 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.748999 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.749022 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.749038 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:12Z","lastTransitionTime":"2025-10-11T03:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.828487 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:12Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.842207 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:12Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.852234 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.852291 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.852306 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.852325 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.852339 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:12Z","lastTransitionTime":"2025-10-11T03:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.857135 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:12Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.871147 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:12Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.890634 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:12Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.903571 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:12Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.919580 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:12Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.951294 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:12Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.954447 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.954603 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.954724 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.954853 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.954962 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:12Z","lastTransitionTime":"2025-10-11T03:52:12Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.966807 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad406c24472c9ef7da82485a0600cf54f9002b83c0770c2ac86eca0f8779a870\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:12Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:12 crc kubenswrapper[4967]: I1011 03:52:12.981480 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"844146aa-36ac-4bee-8015-0d9850d94c4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cafb41d365e674a55638718df4e540ff77b6bd8e0d2d94bdb480b6617c0942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8f8a8bafbf88541375edb7c1b2134dec31a0d5cd46462a3271d24131bed4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hs9vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:12Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.005404 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://257b2303215c18aa4500478fbfa39f9e1b95678c6d39ea51e4324c58f51dbd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.028195 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.045101 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.057784 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.057989 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.058110 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.058233 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.058339 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:13Z","lastTransitionTime":"2025-10-11T03:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.061800 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.082774 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.116145 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3adee99370856f7d1dd95c3fed98740090f99e3d25c449f6f36170f0aeacd3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3adee99370856f7d1dd95c3fed98740090f99e3d25c449f6f36170f0aeacd3e9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:52:00Z\\\",\\\"message\\\":\\\":311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:52:00.856605 6612 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.856659 6612 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.856768 6612 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.857120 6612 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:52:00.857719 6612 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1011 03:52:00.857743 6612 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1011 03:52:00.857756 6612 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1011 03:52:00.857789 6612 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1011 03:52:00.857813 6612 factory.go:656] Stopping watch factory\\\\nI1011 03:52:00.857815 6612 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1011 03:52:00.857835 6612 ovnkube.go:599] Stopped ovnkube\\\\nI1011 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lw6dx_openshift-ovn-kubernetes(270ac8da-b130-4f4f-bbf7-655229b64446)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.136433 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p24x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22c953c-0315-406e-9c5d-f1759837750e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p24x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.158644 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b09975d-4c55-4b1a-b034-851ebf64cd3f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b90baa6dbb24ee4c2deea587bada1820917fd963caaf91210c25409b254c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fef6a4dbe93461c6352741acfc9439a22668f46dd725fcf80fb37d93c88e9d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f23f4ffaeac5c8606def100262484a45cd86296beb91735215c57955e45067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df874a11e199cea6058abe0e5bb63faf7246199fa02093f05720bc41dbb7a81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df874a11e199cea6058abe0e5bb63faf7246199fa02093f05720bc41dbb7a81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:13Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.161294 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.161362 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.161382 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.161409 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.161428 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:13Z","lastTransitionTime":"2025-10-11T03:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.264344 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.264405 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.264421 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.264444 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.264464 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:13Z","lastTransitionTime":"2025-10-11T03:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.366674 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.366746 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.366763 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.366787 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.366804 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:13Z","lastTransitionTime":"2025-10-11T03:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.470010 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.470132 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.470150 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.470175 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.470195 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:13Z","lastTransitionTime":"2025-10-11T03:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.574056 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.574140 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.574198 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.574227 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.574245 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:13Z","lastTransitionTime":"2025-10-11T03:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.677384 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.677460 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.677483 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.677518 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.677542 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:13Z","lastTransitionTime":"2025-10-11T03:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.780477 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.780544 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.780568 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.780599 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.780623 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:13Z","lastTransitionTime":"2025-10-11T03:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.814627 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.814657 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.814758 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.814773 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:13 crc kubenswrapper[4967]: E1011 03:52:13.814931 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:52:13 crc kubenswrapper[4967]: E1011 03:52:13.815194 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:52:13 crc kubenswrapper[4967]: E1011 03:52:13.815346 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:52:13 crc kubenswrapper[4967]: E1011 03:52:13.815555 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.884944 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.885442 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.885674 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.885899 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.886060 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:13Z","lastTransitionTime":"2025-10-11T03:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.989243 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.989278 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.989291 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.989307 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:13 crc kubenswrapper[4967]: I1011 03:52:13.989317 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:13Z","lastTransitionTime":"2025-10-11T03:52:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.092499 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.092550 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.092567 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.092593 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.092621 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:14Z","lastTransitionTime":"2025-10-11T03:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.196166 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.196230 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.196249 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.196273 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.196294 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:14Z","lastTransitionTime":"2025-10-11T03:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.299300 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.299366 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.299389 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.299416 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.299436 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:14Z","lastTransitionTime":"2025-10-11T03:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.403359 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.403420 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.403436 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.403457 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.403480 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:14Z","lastTransitionTime":"2025-10-11T03:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.506528 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.506598 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.506616 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.506643 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.506686 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:14Z","lastTransitionTime":"2025-10-11T03:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.611289 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.611356 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.611374 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.611399 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.611416 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:14Z","lastTransitionTime":"2025-10-11T03:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.714594 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.714975 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.715530 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.715686 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.715817 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:14Z","lastTransitionTime":"2025-10-11T03:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.819391 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.819886 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.820167 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.820388 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.820584 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:14Z","lastTransitionTime":"2025-10-11T03:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.924515 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.924589 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.924612 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.924642 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:14 crc kubenswrapper[4967]: I1011 03:52:14.924661 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:14Z","lastTransitionTime":"2025-10-11T03:52:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.028052 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.028211 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.028237 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.028267 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.028290 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:15Z","lastTransitionTime":"2025-10-11T03:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.132622 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.132689 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.132702 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.132723 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.132735 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:15Z","lastTransitionTime":"2025-10-11T03:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.235191 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.235298 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.235315 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.235340 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.235358 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:15Z","lastTransitionTime":"2025-10-11T03:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.340658 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.340735 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.340782 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.340818 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.340842 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:15Z","lastTransitionTime":"2025-10-11T03:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.444845 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.444924 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.444943 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.444968 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.444984 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:15Z","lastTransitionTime":"2025-10-11T03:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.548112 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.548181 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.548204 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.548232 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.548253 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:15Z","lastTransitionTime":"2025-10-11T03:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.651871 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.651943 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.651965 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.651994 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.652013 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:15Z","lastTransitionTime":"2025-10-11T03:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.754439 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.754564 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.754585 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.754607 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.754623 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:15Z","lastTransitionTime":"2025-10-11T03:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.814409 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.814441 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.814479 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.814479 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:15 crc kubenswrapper[4967]: E1011 03:52:15.814580 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:52:15 crc kubenswrapper[4967]: E1011 03:52:15.814675 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:52:15 crc kubenswrapper[4967]: E1011 03:52:15.814773 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:52:15 crc kubenswrapper[4967]: E1011 03:52:15.815241 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.815489 4967 scope.go:117] "RemoveContainer" containerID="3adee99370856f7d1dd95c3fed98740090f99e3d25c449f6f36170f0aeacd3e9" Oct 11 03:52:15 crc kubenswrapper[4967]: E1011 03:52:15.815741 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lw6dx_openshift-ovn-kubernetes(270ac8da-b130-4f4f-bbf7-655229b64446)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.856392 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.856422 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.856431 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.856443 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.856454 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:15Z","lastTransitionTime":"2025-10-11T03:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.958662 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.958718 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.958734 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.958754 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:15 crc kubenswrapper[4967]: I1011 03:52:15.958768 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:15Z","lastTransitionTime":"2025-10-11T03:52:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.060717 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.060915 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.060927 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.060944 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.060956 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:16Z","lastTransitionTime":"2025-10-11T03:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.163264 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.163305 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.163317 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.163334 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.163346 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:16Z","lastTransitionTime":"2025-10-11T03:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.265842 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.265881 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.265893 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.265907 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.265919 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:16Z","lastTransitionTime":"2025-10-11T03:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.368601 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.368658 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.368675 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.368698 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.368713 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:16Z","lastTransitionTime":"2025-10-11T03:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.471694 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.471758 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.471775 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.471801 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.471818 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:16Z","lastTransitionTime":"2025-10-11T03:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.575422 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.575514 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.575536 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.575562 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.575620 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:16Z","lastTransitionTime":"2025-10-11T03:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.679156 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.679215 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.679235 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.679261 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.679278 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:16Z","lastTransitionTime":"2025-10-11T03:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.782354 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.782412 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.782440 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.782471 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.782513 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:16Z","lastTransitionTime":"2025-10-11T03:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.885841 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.886280 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.886505 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.886664 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.886882 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:16Z","lastTransitionTime":"2025-10-11T03:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.990059 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.990233 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.990258 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.990290 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:16 crc kubenswrapper[4967]: I1011 03:52:16.990310 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:16Z","lastTransitionTime":"2025-10-11T03:52:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.092653 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.092692 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.092700 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.092714 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.092725 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:17Z","lastTransitionTime":"2025-10-11T03:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.195156 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.195218 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.195242 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.195270 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.195292 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:17Z","lastTransitionTime":"2025-10-11T03:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.297563 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.297601 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.297612 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.297628 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.297639 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:17Z","lastTransitionTime":"2025-10-11T03:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.403205 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.403263 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.403282 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.403301 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.403316 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:17Z","lastTransitionTime":"2025-10-11T03:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.505453 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.505509 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.505522 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.505540 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.505550 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:17Z","lastTransitionTime":"2025-10-11T03:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.608444 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.608499 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.608511 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.608527 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.608538 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:17Z","lastTransitionTime":"2025-10-11T03:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.711259 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.711301 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.711310 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.711325 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.711339 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:17Z","lastTransitionTime":"2025-10-11T03:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.813868 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.813905 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.813954 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:17 crc kubenswrapper[4967]: E1011 03:52:17.814039 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.813868 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.814111 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.814155 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.814174 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.814198 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:17 crc kubenswrapper[4967]: E1011 03:52:17.814202 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.814217 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:17Z","lastTransitionTime":"2025-10-11T03:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:17 crc kubenswrapper[4967]: E1011 03:52:17.814298 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:52:17 crc kubenswrapper[4967]: E1011 03:52:17.814378 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.916817 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.916844 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.916853 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.916865 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:17 crc kubenswrapper[4967]: I1011 03:52:17.916875 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:17Z","lastTransitionTime":"2025-10-11T03:52:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.019224 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.019271 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.019284 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.019302 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.019313 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:18Z","lastTransitionTime":"2025-10-11T03:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.121709 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.122199 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.122264 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.122338 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.122404 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:18Z","lastTransitionTime":"2025-10-11T03:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.225068 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.225295 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.225384 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.225460 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.225539 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:18Z","lastTransitionTime":"2025-10-11T03:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.327980 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.328028 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.328039 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.328058 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.328093 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:18Z","lastTransitionTime":"2025-10-11T03:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.430207 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.430244 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.430255 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.430272 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.430285 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:18Z","lastTransitionTime":"2025-10-11T03:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.532292 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.532344 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.532359 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.532378 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.532393 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:18Z","lastTransitionTime":"2025-10-11T03:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.635510 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.635555 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.635566 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.635582 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.635595 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:18Z","lastTransitionTime":"2025-10-11T03:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.737989 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.738046 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.738063 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.738127 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.738150 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:18Z","lastTransitionTime":"2025-10-11T03:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.840042 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.840125 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.840142 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.840163 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.840183 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:18Z","lastTransitionTime":"2025-10-11T03:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.942834 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.942895 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.942918 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.942945 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:18 crc kubenswrapper[4967]: I1011 03:52:18.942967 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:18Z","lastTransitionTime":"2025-10-11T03:52:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.044976 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.045003 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.045011 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.045025 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.045034 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:19Z","lastTransitionTime":"2025-10-11T03:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.147722 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.147826 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.147852 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.147882 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.147908 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:19Z","lastTransitionTime":"2025-10-11T03:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.250421 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.250474 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.250492 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.250513 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.250530 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:19Z","lastTransitionTime":"2025-10-11T03:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.353418 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.353476 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.353495 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.353517 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.353534 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:19Z","lastTransitionTime":"2025-10-11T03:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.450745 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.450776 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.450784 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.450794 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.450802 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:19Z","lastTransitionTime":"2025-10-11T03:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:19 crc kubenswrapper[4967]: E1011 03:52:19.462524 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.467265 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.467287 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.467295 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.467305 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.467312 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:19Z","lastTransitionTime":"2025-10-11T03:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:19 crc kubenswrapper[4967]: E1011 03:52:19.477737 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.482065 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.482106 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.482114 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.482125 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.482134 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:19Z","lastTransitionTime":"2025-10-11T03:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:19 crc kubenswrapper[4967]: E1011 03:52:19.497953 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.501925 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.501989 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.502003 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.502019 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.502031 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:19Z","lastTransitionTime":"2025-10-11T03:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:19 crc kubenswrapper[4967]: E1011 03:52:19.518151 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.522572 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.522596 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.522623 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.522634 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.522643 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:19Z","lastTransitionTime":"2025-10-11T03:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:19 crc kubenswrapper[4967]: E1011 03:52:19.538310 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:19Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:19Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:19 crc kubenswrapper[4967]: E1011 03:52:19.538536 4967 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.540578 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.540746 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.540767 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.540790 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.540809 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:19Z","lastTransitionTime":"2025-10-11T03:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.642607 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.642684 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.642701 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.642722 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.642737 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:19Z","lastTransitionTime":"2025-10-11T03:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.747634 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.748352 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.748781 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.749020 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.749281 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:19Z","lastTransitionTime":"2025-10-11T03:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.814518 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.814548 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.814559 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.814608 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:19 crc kubenswrapper[4967]: E1011 03:52:19.814705 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:52:19 crc kubenswrapper[4967]: E1011 03:52:19.814805 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:52:19 crc kubenswrapper[4967]: E1011 03:52:19.814909 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:52:19 crc kubenswrapper[4967]: E1011 03:52:19.815068 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.851434 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.851641 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.851732 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.851819 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.851916 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:19Z","lastTransitionTime":"2025-10-11T03:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.954313 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.954369 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.954387 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.954412 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.954430 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:19Z","lastTransitionTime":"2025-10-11T03:52:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:19 crc kubenswrapper[4967]: I1011 03:52:19.988735 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d22c953c-0315-406e-9c5d-f1759837750e-metrics-certs\") pod \"network-metrics-daemon-p24x8\" (UID: \"d22c953c-0315-406e-9c5d-f1759837750e\") " pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:19 crc kubenswrapper[4967]: E1011 03:52:19.988886 4967 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 11 03:52:19 crc kubenswrapper[4967]: E1011 03:52:19.988988 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d22c953c-0315-406e-9c5d-f1759837750e-metrics-certs podName:d22c953c-0315-406e-9c5d-f1759837750e nodeName:}" failed. No retries permitted until 2025-10-11 03:52:51.98896414 +0000 UTC m=+99.952173113 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d22c953c-0315-406e-9c5d-f1759837750e-metrics-certs") pod "network-metrics-daemon-p24x8" (UID: "d22c953c-0315-406e-9c5d-f1759837750e") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.056554 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.056594 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.056602 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.056617 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.056630 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:20Z","lastTransitionTime":"2025-10-11T03:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.159612 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.159657 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.159668 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.159684 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.159696 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:20Z","lastTransitionTime":"2025-10-11T03:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.262191 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.262248 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.262261 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.262275 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.262287 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:20Z","lastTransitionTime":"2025-10-11T03:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.365138 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.365481 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.365627 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.365773 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.365905 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:20Z","lastTransitionTime":"2025-10-11T03:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.468690 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.468729 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.468738 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.468752 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.468761 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:20Z","lastTransitionTime":"2025-10-11T03:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.571111 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.571171 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.571187 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.571212 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.571228 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:20Z","lastTransitionTime":"2025-10-11T03:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.673527 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.673575 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.673591 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.673613 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.673629 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:20Z","lastTransitionTime":"2025-10-11T03:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.775609 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.775664 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.775680 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.775705 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.775725 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:20Z","lastTransitionTime":"2025-10-11T03:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.877796 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.877870 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.877884 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.877902 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.877916 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:20Z","lastTransitionTime":"2025-10-11T03:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.979999 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.980102 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.980129 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.980156 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:20 crc kubenswrapper[4967]: I1011 03:52:20.980175 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:20Z","lastTransitionTime":"2025-10-11T03:52:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.083418 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.083479 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.083496 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.083519 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.083536 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:21Z","lastTransitionTime":"2025-10-11T03:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.186094 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.186129 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.186138 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.186157 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.186167 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:21Z","lastTransitionTime":"2025-10-11T03:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.228850 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6xvgx_1db525c3-0479-4585-b107-a448f2b94ffd/kube-multus/0.log" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.228893 4967 generic.go:334] "Generic (PLEG): container finished" podID="1db525c3-0479-4585-b107-a448f2b94ffd" containerID="a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0" exitCode=1 Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.228918 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6xvgx" event={"ID":"1db525c3-0479-4585-b107-a448f2b94ffd","Type":"ContainerDied","Data":"a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0"} Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.229246 4967 scope.go:117] "RemoveContainer" containerID="a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.242149 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:21Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.256008 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:21Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.268338 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:21Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.280908 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:21Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.291513 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.291571 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.291582 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.291597 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.291607 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:21Z","lastTransitionTime":"2025-10-11T03:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.294330 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b09975d-4c55-4b1a-b034-851ebf64cd3f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b90baa6dbb24ee4c2deea587bada1820917fd963caaf91210c25409b254c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fef6a4dbe93461c6352741acfc9439a22668f46dd725fcf80fb37d93c88e9d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f23f4ffaeac5c8606def100262484a45cd86296beb91735215c57955e45067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df874a11e199cea6058abe0e5bb63faf7246199fa02093f05720bc41dbb7a81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df874a11e199cea6058abe0e5bb63faf7246199fa02093f05720bc41dbb7a81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:21Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.312356 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3adee99370856f7d1dd95c3fed98740090f99e3d25c449f6f36170f0aeacd3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3adee99370856f7d1dd95c3fed98740090f99e3d25c449f6f36170f0aeacd3e9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:52:00Z\\\",\\\"message\\\":\\\":311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:52:00.856605 6612 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.856659 6612 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.856768 6612 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.857120 6612 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:52:00.857719 6612 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1011 03:52:00.857743 6612 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1011 03:52:00.857756 6612 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1011 03:52:00.857789 6612 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1011 03:52:00.857813 6612 factory.go:656] Stopping watch factory\\\\nI1011 03:52:00.857815 6612 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1011 03:52:00.857835 6612 ovnkube.go:599] Stopped ovnkube\\\\nI1011 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lw6dx_openshift-ovn-kubernetes(270ac8da-b130-4f4f-bbf7-655229b64446)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:21Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.326201 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p24x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22c953c-0315-406e-9c5d-f1759837750e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p24x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:21Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.343200 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:21Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.367171 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:21Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.381391 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:21Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.392201 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:21Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.394019 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.394058 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.394099 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.394121 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.394136 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:21Z","lastTransitionTime":"2025-10-11T03:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.407349 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:21Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:52:21Z\\\",\\\"message\\\":\\\"2025-10-11T03:51:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_74aecb9d-f601-43f8-b48e-632d0155dabc\\\\n2025-10-11T03:51:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_74aecb9d-f601-43f8-b48e-632d0155dabc to /host/opt/cni/bin/\\\\n2025-10-11T03:51:36Z [verbose] multus-daemon started\\\\n2025-10-11T03:51:36Z [verbose] Readiness Indicator file check\\\\n2025-10-11T03:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:21Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.429769 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:21Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.442787 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:21Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.457117 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:21Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.477756 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://257b2303215c18aa4500478fbfa39f9e1b95678c6d39ea51e4324c58f51dbd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:21Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.491473 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad406c24472c9ef7da82485a0600cf54f9002b83c0770c2ac86eca0f8779a870\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:21Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.495873 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.495908 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.495917 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.495931 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.495942 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:21Z","lastTransitionTime":"2025-10-11T03:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.508433 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"844146aa-36ac-4bee-8015-0d9850d94c4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cafb41d365e674a55638718df4e540ff77b6bd8e0d2d94bdb480b6617c0942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8f8a8bafbf88541375edb7c1b2134dec31a0d5cd46462a3271d24131bed4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hs9vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:21Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.598682 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.598722 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.598732 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.598747 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.598757 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:21Z","lastTransitionTime":"2025-10-11T03:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.701313 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.701354 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.701362 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.701377 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.701387 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:21Z","lastTransitionTime":"2025-10-11T03:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.804359 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.804404 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.804413 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.804429 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.804439 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:21Z","lastTransitionTime":"2025-10-11T03:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.813770 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.813798 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.813839 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:52:21 crc kubenswrapper[4967]: E1011 03:52:21.813896 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.813780 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:52:21 crc kubenswrapper[4967]: E1011 03:52:21.814118 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:52:21 crc kubenswrapper[4967]: E1011 03:52:21.814166 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:52:21 crc kubenswrapper[4967]: E1011 03:52:21.814284 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.906595 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.906641 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.906649 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.906663 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:21 crc kubenswrapper[4967]: I1011 03:52:21.906672 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:21Z","lastTransitionTime":"2025-10-11T03:52:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.009332 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.009379 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.009388 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.009402 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.009411 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:22Z","lastTransitionTime":"2025-10-11T03:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.111931 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.111975 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.111984 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.111998 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.112007 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:22Z","lastTransitionTime":"2025-10-11T03:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.214419 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.214459 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.214469 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.214483 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.214494 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:22Z","lastTransitionTime":"2025-10-11T03:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.233112 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6xvgx_1db525c3-0479-4585-b107-a448f2b94ffd/kube-multus/0.log" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.233151 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6xvgx" event={"ID":"1db525c3-0479-4585-b107-a448f2b94ffd","Type":"ContainerStarted","Data":"83b1e937441999be327ba6a58c849e9c04956f5cea84da9f8f4aa476e1e79c76"} Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.253725 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.271346 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.289399 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.304418 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.317619 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.317657 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.317673 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.317695 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.317711 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:22Z","lastTransitionTime":"2025-10-11T03:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.322803 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83b1e937441999be327ba6a58c849e9c04956f5cea84da9f8f4aa476e1e79c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:52:21Z\\\",\\\"message\\\":\\\"2025-10-11T03:51:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_74aecb9d-f601-43f8-b48e-632d0155dabc\\\\n2025-10-11T03:51:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_74aecb9d-f601-43f8-b48e-632d0155dabc to /host/opt/cni/bin/\\\\n2025-10-11T03:51:36Z [verbose] multus-daemon started\\\\n2025-10-11T03:51:36Z [verbose] Readiness Indicator file check\\\\n2025-10-11T03:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.354659 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.373254 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.392653 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.415553 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://257b2303215c18aa4500478fbfa39f9e1b95678c6d39ea51e4324c58f51dbd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.419990 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.420132 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.420209 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.420329 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.420427 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:22Z","lastTransitionTime":"2025-10-11T03:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.430676 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad406c24472c9ef7da82485a0600cf54f9002b83c0770c2ac86eca0f8779a870\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.445572 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"844146aa-36ac-4bee-8015-0d9850d94c4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cafb41d365e674a55638718df4e540ff77b6bd8e0d2d94bdb480b6617c0942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8f8a8bafbf88541375edb7c1b2134dec31a0d5cd46462a3271d24131bed4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hs9vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.461918 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.481522 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.494204 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.506013 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.515569 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b09975d-4c55-4b1a-b034-851ebf64cd3f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b90baa6dbb24ee4c2deea587bada1820917fd963caaf91210c25409b254c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fef6a4dbe93461c6352741acfc9439a22668f46dd725fcf80fb37d93c88e9d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f23f4ffaeac5c8606def100262484a45cd86296beb91735215c57955e45067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df874a11e199cea6058abe0e5bb63faf7246199fa02093f05720bc41dbb7a81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df874a11e199cea6058abe0e5bb63faf7246199fa02093f05720bc41dbb7a81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.522504 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.522533 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.522544 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.522561 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.522574 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:22Z","lastTransitionTime":"2025-10-11T03:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.533055 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3adee99370856f7d1dd95c3fed98740090f99e3d25c449f6f36170f0aeacd3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3adee99370856f7d1dd95c3fed98740090f99e3d25c449f6f36170f0aeacd3e9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:52:00Z\\\",\\\"message\\\":\\\":311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:52:00.856605 6612 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.856659 6612 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.856768 6612 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.857120 6612 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:52:00.857719 6612 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1011 03:52:00.857743 6612 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1011 03:52:00.857756 6612 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1011 03:52:00.857789 6612 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1011 03:52:00.857813 6612 factory.go:656] Stopping watch factory\\\\nI1011 03:52:00.857815 6612 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1011 03:52:00.857835 6612 ovnkube.go:599] Stopped ovnkube\\\\nI1011 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lw6dx_openshift-ovn-kubernetes(270ac8da-b130-4f4f-bbf7-655229b64446)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.542319 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p24x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22c953c-0315-406e-9c5d-f1759837750e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p24x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.625258 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.625288 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.625322 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.625338 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.625350 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:22Z","lastTransitionTime":"2025-10-11T03:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.727729 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.728092 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.728183 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.728282 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.728388 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:22Z","lastTransitionTime":"2025-10-11T03:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.829983 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.830012 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.830022 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.830034 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.830043 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:22Z","lastTransitionTime":"2025-10-11T03:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.831134 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83b1e937441999be327ba6a58c849e9c04956f5cea84da9f8f4aa476e1e79c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:52:21Z\\\",\\\"message\\\":\\\"2025-10-11T03:51:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_74aecb9d-f601-43f8-b48e-632d0155dabc\\\\n2025-10-11T03:51:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_74aecb9d-f601-43f8-b48e-632d0155dabc to /host/opt/cni/bin/\\\\n2025-10-11T03:51:36Z [verbose] multus-daemon started\\\\n2025-10-11T03:51:36Z [verbose] Readiness Indicator file check\\\\n2025-10-11T03:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.851888 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.863612 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.874857 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.892345 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.906377 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.924520 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.933601 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.933656 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.933667 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.933683 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.933694 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:22Z","lastTransitionTime":"2025-10-11T03:52:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.935557 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.949482 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://257b2303215c18aa4500478fbfa39f9e1b95678c6d39ea51e4324c58f51dbd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.960040 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad406c24472c9ef7da82485a0600cf54f9002b83c0770c2ac86eca0f8779a870\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.970246 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"844146aa-36ac-4bee-8015-0d9850d94c4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cafb41d365e674a55638718df4e540ff77b6bd8e0d2d94bdb480b6617c0942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8f8a8bafbf88541375edb7c1b2134dec31a0d5cd46462a3271d24131bed4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hs9vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.981984 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:22 crc kubenswrapper[4967]: I1011 03:52:22.991697 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:22Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.005043 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.014136 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.027105 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b09975d-4c55-4b1a-b034-851ebf64cd3f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b90baa6dbb24ee4c2deea587bada1820917fd963caaf91210c25409b254c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fef6a4dbe93461c6352741acfc9439a22668f46dd725fcf80fb37d93c88e9d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f23f4ffaeac5c8606def100262484a45cd86296beb91735215c57955e45067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df874a11e199cea6058abe0e5bb63faf7246199fa02093f05720bc41dbb7a81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df874a11e199cea6058abe0e5bb63faf7246199fa02093f05720bc41dbb7a81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.036106 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.036136 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.036147 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.036183 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.036194 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:23Z","lastTransitionTime":"2025-10-11T03:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.046009 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://3adee99370856f7d1dd95c3fed98740090f99e3d25c449f6f36170f0aeacd3e9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3adee99370856f7d1dd95c3fed98740090f99e3d25c449f6f36170f0aeacd3e9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:52:00Z\\\",\\\"message\\\":\\\":311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:52:00.856605 6612 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.856659 6612 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.856768 6612 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.857120 6612 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:52:00.857719 6612 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1011 03:52:00.857743 6612 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1011 03:52:00.857756 6612 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1011 03:52:00.857789 6612 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1011 03:52:00.857813 6612 factory.go:656] Stopping watch factory\\\\nI1011 03:52:00.857815 6612 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1011 03:52:00.857835 6612 ovnkube.go:599] Stopped ovnkube\\\\nI1011 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-lw6dx_openshift-ovn-kubernetes(270ac8da-b130-4f4f-bbf7-655229b64446)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.055174 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p24x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22c953c-0315-406e-9c5d-f1759837750e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p24x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:23Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.138961 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.139004 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.139013 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.139028 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.139039 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:23Z","lastTransitionTime":"2025-10-11T03:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.240926 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.240967 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.240976 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.240990 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.241001 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:23Z","lastTransitionTime":"2025-10-11T03:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.345123 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.345184 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.345202 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.345228 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.345245 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:23Z","lastTransitionTime":"2025-10-11T03:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.447506 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.447572 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.447591 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.447616 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.447633 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:23Z","lastTransitionTime":"2025-10-11T03:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.550199 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.550241 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.550252 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.550270 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.550282 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:23Z","lastTransitionTime":"2025-10-11T03:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.653482 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.653533 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.653552 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.653574 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.653590 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:23Z","lastTransitionTime":"2025-10-11T03:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.755560 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.755745 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.755822 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.755883 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.755943 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:23Z","lastTransitionTime":"2025-10-11T03:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.814254 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:52:23 crc kubenswrapper[4967]: E1011 03:52:23.814520 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.814310 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:52:23 crc kubenswrapper[4967]: E1011 03:52:23.814835 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.814254 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:23 crc kubenswrapper[4967]: E1011 03:52:23.815273 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.814385 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:23 crc kubenswrapper[4967]: E1011 03:52:23.815598 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.858449 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.858662 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.858772 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.858874 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.858974 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:23Z","lastTransitionTime":"2025-10-11T03:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.961259 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.961326 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.961349 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.961378 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:23 crc kubenswrapper[4967]: I1011 03:52:23.961409 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:23Z","lastTransitionTime":"2025-10-11T03:52:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.064067 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.064116 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.064142 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.064157 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.064166 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:24Z","lastTransitionTime":"2025-10-11T03:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.166378 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.166434 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.166455 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.166486 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.166509 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:24Z","lastTransitionTime":"2025-10-11T03:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.299230 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.299296 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.299313 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.299342 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.299366 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:24Z","lastTransitionTime":"2025-10-11T03:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.401242 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.401505 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.401586 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.401683 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.401773 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:24Z","lastTransitionTime":"2025-10-11T03:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.504387 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.504443 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.504459 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.504481 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.504497 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:24Z","lastTransitionTime":"2025-10-11T03:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.607474 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.607514 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.607525 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.607541 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.607554 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:24Z","lastTransitionTime":"2025-10-11T03:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.710700 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.710740 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.710749 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.710765 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.710776 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:24Z","lastTransitionTime":"2025-10-11T03:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.812878 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.812921 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.812931 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.812945 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.812955 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:24Z","lastTransitionTime":"2025-10-11T03:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.916646 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.916710 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.916727 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.916755 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:24 crc kubenswrapper[4967]: I1011 03:52:24.916773 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:24Z","lastTransitionTime":"2025-10-11T03:52:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.019114 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.019149 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.019160 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.019174 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.019187 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:25Z","lastTransitionTime":"2025-10-11T03:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.122053 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.122119 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.122133 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.122150 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.122162 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:25Z","lastTransitionTime":"2025-10-11T03:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.224705 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.224763 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.224782 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.224803 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.224818 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:25Z","lastTransitionTime":"2025-10-11T03:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.327238 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.327367 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.327386 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.327411 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.327429 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:25Z","lastTransitionTime":"2025-10-11T03:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.430436 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.430496 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.430516 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.430545 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.430571 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:25Z","lastTransitionTime":"2025-10-11T03:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.533130 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.533338 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.533397 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.533485 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.533552 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:25Z","lastTransitionTime":"2025-10-11T03:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.636053 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.636447 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.636574 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.636684 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.636772 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:25Z","lastTransitionTime":"2025-10-11T03:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.740259 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.740501 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.740562 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.740627 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.740721 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:25Z","lastTransitionTime":"2025-10-11T03:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.813904 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.814014 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:52:25 crc kubenswrapper[4967]: E1011 03:52:25.814309 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.814097 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:25 crc kubenswrapper[4967]: E1011 03:52:25.814632 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.814038 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:25 crc kubenswrapper[4967]: E1011 03:52:25.814886 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:52:25 crc kubenswrapper[4967]: E1011 03:52:25.814429 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.843786 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.843847 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.843867 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.843920 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.843940 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:25Z","lastTransitionTime":"2025-10-11T03:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.948594 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.948662 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.948680 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.948759 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:25 crc kubenswrapper[4967]: I1011 03:52:25.948780 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:25Z","lastTransitionTime":"2025-10-11T03:52:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.050933 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.050975 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.050991 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.051013 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.051029 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:26Z","lastTransitionTime":"2025-10-11T03:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.153811 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.153872 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.153891 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.153975 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.153993 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:26Z","lastTransitionTime":"2025-10-11T03:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.256388 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.256653 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.256729 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.256789 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.256849 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:26Z","lastTransitionTime":"2025-10-11T03:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.359510 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.359986 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.360264 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.360366 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.360451 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:26Z","lastTransitionTime":"2025-10-11T03:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.462582 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.462614 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.462625 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.462640 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.462652 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:26Z","lastTransitionTime":"2025-10-11T03:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.565440 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.565509 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.565531 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.565559 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.565580 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:26Z","lastTransitionTime":"2025-10-11T03:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.668619 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.668692 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.668719 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.668747 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.668767 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:26Z","lastTransitionTime":"2025-10-11T03:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.770879 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.770914 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.770925 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.770941 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.770952 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:26Z","lastTransitionTime":"2025-10-11T03:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.873449 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.873484 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.873494 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.873509 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.873518 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:26Z","lastTransitionTime":"2025-10-11T03:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.976639 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.976674 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.976683 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.976697 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:26 crc kubenswrapper[4967]: I1011 03:52:26.976705 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:26Z","lastTransitionTime":"2025-10-11T03:52:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.079900 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.080868 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.080890 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.080916 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.080935 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:27Z","lastTransitionTime":"2025-10-11T03:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.183664 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.183764 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.183788 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.183864 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.183886 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:27Z","lastTransitionTime":"2025-10-11T03:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.287011 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.287067 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.287126 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.287150 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.287167 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:27Z","lastTransitionTime":"2025-10-11T03:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.390123 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.390151 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.390159 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.390171 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.390179 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:27Z","lastTransitionTime":"2025-10-11T03:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.492565 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.492611 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.492623 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.492643 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.492656 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:27Z","lastTransitionTime":"2025-10-11T03:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.596116 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.596157 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.596168 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.596185 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.596198 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:27Z","lastTransitionTime":"2025-10-11T03:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.698481 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.698528 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.698539 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.698556 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.698570 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:27Z","lastTransitionTime":"2025-10-11T03:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.801772 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.801834 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.801851 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.801877 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.801894 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:27Z","lastTransitionTime":"2025-10-11T03:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.814197 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.814225 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.814235 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.814311 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:52:27 crc kubenswrapper[4967]: E1011 03:52:27.814501 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:52:27 crc kubenswrapper[4967]: E1011 03:52:27.814631 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:52:27 crc kubenswrapper[4967]: E1011 03:52:27.814782 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:52:27 crc kubenswrapper[4967]: E1011 03:52:27.814875 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.904748 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.904822 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.904838 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.904863 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:27 crc kubenswrapper[4967]: I1011 03:52:27.904881 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:27Z","lastTransitionTime":"2025-10-11T03:52:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.007983 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.008053 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.008094 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.008120 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.008146 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:28Z","lastTransitionTime":"2025-10-11T03:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.110890 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.110967 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.111007 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.111037 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.111057 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:28Z","lastTransitionTime":"2025-10-11T03:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.214250 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.214311 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.214327 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.214350 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.214367 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:28Z","lastTransitionTime":"2025-10-11T03:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.317060 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.317209 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.317228 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.317251 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.317270 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:28Z","lastTransitionTime":"2025-10-11T03:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.429915 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.429982 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.430002 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.430028 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.430047 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:28Z","lastTransitionTime":"2025-10-11T03:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.532457 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.532489 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.532500 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.532517 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.532528 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:28Z","lastTransitionTime":"2025-10-11T03:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.636007 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.636525 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.636694 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.636943 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.637137 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:28Z","lastTransitionTime":"2025-10-11T03:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.739673 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.739741 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.739754 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.739770 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.739780 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:28Z","lastTransitionTime":"2025-10-11T03:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.842941 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.843025 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.843050 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.843113 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.843134 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:28Z","lastTransitionTime":"2025-10-11T03:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.946616 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.946661 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.946675 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.946690 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:28 crc kubenswrapper[4967]: I1011 03:52:28.946702 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:28Z","lastTransitionTime":"2025-10-11T03:52:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.049537 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.052228 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.052307 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.052337 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.052362 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:29Z","lastTransitionTime":"2025-10-11T03:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.156621 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.156695 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.156717 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.156751 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.156774 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:29Z","lastTransitionTime":"2025-10-11T03:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.259919 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.259960 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.259971 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.259989 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.260002 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:29Z","lastTransitionTime":"2025-10-11T03:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.363320 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.363390 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.363411 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.363437 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.363455 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:29Z","lastTransitionTime":"2025-10-11T03:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.466890 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.466954 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.466971 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.466995 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.467015 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:29Z","lastTransitionTime":"2025-10-11T03:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.569917 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.570002 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.570022 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.570049 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.570067 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:29Z","lastTransitionTime":"2025-10-11T03:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.673235 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.673315 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.673339 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.673369 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.673387 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:29Z","lastTransitionTime":"2025-10-11T03:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.745962 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.746045 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.746066 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.746122 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.746143 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:29Z","lastTransitionTime":"2025-10-11T03:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:29 crc kubenswrapper[4967]: E1011 03:52:29.761389 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:29Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.767826 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.767870 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.767881 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.767903 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.767917 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:29Z","lastTransitionTime":"2025-10-11T03:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:29 crc kubenswrapper[4967]: E1011 03:52:29.780857 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:29Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.786516 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.786618 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.786636 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.786664 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.786683 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:29Z","lastTransitionTime":"2025-10-11T03:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:29 crc kubenswrapper[4967]: E1011 03:52:29.804062 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:29Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.808954 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.809021 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.809047 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.809124 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.809154 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:29Z","lastTransitionTime":"2025-10-11T03:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.814633 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.814677 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.814633 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:29 crc kubenswrapper[4967]: E1011 03:52:29.814836 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.814910 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:52:29 crc kubenswrapper[4967]: E1011 03:52:29.815059 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:52:29 crc kubenswrapper[4967]: E1011 03:52:29.815309 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:52:29 crc kubenswrapper[4967]: E1011 03:52:29.815482 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:52:29 crc kubenswrapper[4967]: E1011 03:52:29.834226 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:29Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.839024 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.839135 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.839160 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.839184 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.839204 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:29Z","lastTransitionTime":"2025-10-11T03:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:29 crc kubenswrapper[4967]: E1011 03:52:29.860451 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:29Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:29Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:29 crc kubenswrapper[4967]: E1011 03:52:29.860699 4967 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.863544 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.863641 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.863694 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.863724 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.863780 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:29Z","lastTransitionTime":"2025-10-11T03:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.967445 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.967527 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.967551 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.967585 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:29 crc kubenswrapper[4967]: I1011 03:52:29.967608 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:29Z","lastTransitionTime":"2025-10-11T03:52:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.071315 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.071391 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.071414 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.071440 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.071461 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:30Z","lastTransitionTime":"2025-10-11T03:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.174934 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.175503 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.175767 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.175979 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.176219 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:30Z","lastTransitionTime":"2025-10-11T03:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.280402 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.280472 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.280497 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.280533 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.280558 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:30Z","lastTransitionTime":"2025-10-11T03:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.383629 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.383994 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.384253 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.384533 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.384754 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:30Z","lastTransitionTime":"2025-10-11T03:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.487191 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.487240 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.487258 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.487280 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.487297 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:30Z","lastTransitionTime":"2025-10-11T03:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.590656 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.590725 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.590747 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.590776 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.590796 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:30Z","lastTransitionTime":"2025-10-11T03:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.694138 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.694195 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.694210 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.694230 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.694244 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:30Z","lastTransitionTime":"2025-10-11T03:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.798039 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.798385 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.798479 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.798579 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.798665 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:30Z","lastTransitionTime":"2025-10-11T03:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.815400 4967 scope.go:117] "RemoveContainer" containerID="3adee99370856f7d1dd95c3fed98740090f99e3d25c449f6f36170f0aeacd3e9" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.902059 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.902159 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.902184 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.902217 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:30 crc kubenswrapper[4967]: I1011 03:52:30.902239 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:30Z","lastTransitionTime":"2025-10-11T03:52:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.007850 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.008387 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.008452 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.008573 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.008599 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:31Z","lastTransitionTime":"2025-10-11T03:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.111872 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.111904 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.111915 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.111931 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.111943 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:31Z","lastTransitionTime":"2025-10-11T03:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.215993 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.216034 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.216045 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.216093 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.216106 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:31Z","lastTransitionTime":"2025-10-11T03:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.272191 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lw6dx_270ac8da-b130-4f4f-bbf7-655229b64446/ovnkube-controller/2.log" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.276435 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" event={"ID":"270ac8da-b130-4f4f-bbf7-655229b64446","Type":"ContainerStarted","Data":"2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d"} Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.277342 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.301778 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:31Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.319502 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.319551 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.319563 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.319581 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.319593 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:31Z","lastTransitionTime":"2025-10-11T03:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.330959 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:31Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.358694 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:31Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.385694 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:31Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.408279 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b09975d-4c55-4b1a-b034-851ebf64cd3f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b90baa6dbb24ee4c2deea587bada1820917fd963caaf91210c25409b254c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fef6a4dbe93461c6352741acfc9439a22668f46dd725fcf80fb37d93c88e9d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f23f4ffaeac5c8606def100262484a45cd86296beb91735215c57955e45067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df874a11e199cea6058abe0e5bb63faf7246199fa02093f05720bc41dbb7a81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df874a11e199cea6058abe0e5bb63faf7246199fa02093f05720bc41dbb7a81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:31Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.421814 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.421850 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.421864 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.421884 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.421898 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:31Z","lastTransitionTime":"2025-10-11T03:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.426697 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3adee99370856f7d1dd95c3fed98740090f99e3d25c449f6f36170f0aeacd3e9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:52:00Z\\\",\\\"message\\\":\\\":311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:52:00.856605 6612 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.856659 6612 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.856768 6612 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.857120 6612 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:52:00.857719 6612 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1011 03:52:00.857743 6612 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1011 03:52:00.857756 6612 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1011 03:52:00.857789 6612 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1011 03:52:00.857813 6612 factory.go:656] Stopping watch factory\\\\nI1011 03:52:00.857815 6612 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1011 03:52:00.857835 6612 ovnkube.go:599] Stopped ovnkube\\\\nI1011 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:52:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:31Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.439243 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p24x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22c953c-0315-406e-9c5d-f1759837750e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p24x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:31Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.470443 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:31Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.483706 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:31Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.500721 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:31Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.513112 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:31Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.524641 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.524693 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.524704 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.524722 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.524733 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:31Z","lastTransitionTime":"2025-10-11T03:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.528387 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:31Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.547630 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:31Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.561085 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:31Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.574399 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83b1e937441999be327ba6a58c849e9c04956f5cea84da9f8f4aa476e1e79c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:52:21Z\\\",\\\"message\\\":\\\"2025-10-11T03:51:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_74aecb9d-f601-43f8-b48e-632d0155dabc\\\\n2025-10-11T03:51:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_74aecb9d-f601-43f8-b48e-632d0155dabc to /host/opt/cni/bin/\\\\n2025-10-11T03:51:36Z [verbose] multus-daemon started\\\\n2025-10-11T03:51:36Z [verbose] Readiness Indicator file check\\\\n2025-10-11T03:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:31Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.594696 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://257b2303215c18aa4500478fbfa39f9e1b95678c6d39ea51e4324c58f51dbd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:31Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.608390 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad406c24472c9ef7da82485a0600cf54f9002b83c0770c2ac86eca0f8779a870\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:31Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.622493 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"844146aa-36ac-4bee-8015-0d9850d94c4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cafb41d365e674a55638718df4e540ff77b6bd8e0d2d94bdb480b6617c0942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8f8a8bafbf88541375edb7c1b2134dec31a0d5cd46462a3271d24131bed4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hs9vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:31Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.627332 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.627384 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.627396 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.627414 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.627426 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:31Z","lastTransitionTime":"2025-10-11T03:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.729504 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.729550 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.729562 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.729581 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.729593 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:31Z","lastTransitionTime":"2025-10-11T03:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.814673 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.814702 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.814717 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.814688 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:31 crc kubenswrapper[4967]: E1011 03:52:31.814835 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:52:31 crc kubenswrapper[4967]: E1011 03:52:31.814920 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:52:31 crc kubenswrapper[4967]: E1011 03:52:31.814986 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:52:31 crc kubenswrapper[4967]: E1011 03:52:31.815028 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.832446 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.832495 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.832510 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.832532 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.832547 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:31Z","lastTransitionTime":"2025-10-11T03:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.934718 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.934762 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.934776 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.934793 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:31 crc kubenswrapper[4967]: I1011 03:52:31.934806 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:31Z","lastTransitionTime":"2025-10-11T03:52:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.037647 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.037737 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.037761 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.037786 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.037809 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:32Z","lastTransitionTime":"2025-10-11T03:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.141154 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.141217 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.141234 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.141288 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.141310 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:32Z","lastTransitionTime":"2025-10-11T03:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.243967 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.244031 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.244048 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.244104 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.244125 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:32Z","lastTransitionTime":"2025-10-11T03:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.282494 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lw6dx_270ac8da-b130-4f4f-bbf7-655229b64446/ovnkube-controller/3.log" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.283348 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lw6dx_270ac8da-b130-4f4f-bbf7-655229b64446/ovnkube-controller/2.log" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.287370 4967 generic.go:334] "Generic (PLEG): container finished" podID="270ac8da-b130-4f4f-bbf7-655229b64446" containerID="2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d" exitCode=1 Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.287407 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" event={"ID":"270ac8da-b130-4f4f-bbf7-655229b64446","Type":"ContainerDied","Data":"2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d"} Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.287447 4967 scope.go:117] "RemoveContainer" containerID="3adee99370856f7d1dd95c3fed98740090f99e3d25c449f6f36170f0aeacd3e9" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.289560 4967 scope.go:117] "RemoveContainer" containerID="2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d" Oct 11 03:52:32 crc kubenswrapper[4967]: E1011 03:52:32.290010 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lw6dx_openshift-ovn-kubernetes(270ac8da-b130-4f4f-bbf7-655229b64446)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.315215 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.337417 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.346820 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.346880 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.346897 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.346920 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.346938 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:32Z","lastTransitionTime":"2025-10-11T03:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.368183 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.389663 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.409173 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b09975d-4c55-4b1a-b034-851ebf64cd3f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b90baa6dbb24ee4c2deea587bada1820917fd963caaf91210c25409b254c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fef6a4dbe93461c6352741acfc9439a22668f46dd725fcf80fb37d93c88e9d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f23f4ffaeac5c8606def100262484a45cd86296beb91735215c57955e45067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df874a11e199cea6058abe0e5bb63faf7246199fa02093f05720bc41dbb7a81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df874a11e199cea6058abe0e5bb63faf7246199fa02093f05720bc41dbb7a81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.480057 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.480153 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.480171 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.480198 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.480217 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:32Z","lastTransitionTime":"2025-10-11T03:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.482065 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3adee99370856f7d1dd95c3fed98740090f99e3d25c449f6f36170f0aeacd3e9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:52:00Z\\\",\\\"message\\\":\\\":311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:52:00.856605 6612 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.856659 6612 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.856768 6612 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.857120 6612 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:52:00.857719 6612 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1011 03:52:00.857743 6612 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1011 03:52:00.857756 6612 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1011 03:52:00.857789 6612 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1011 03:52:00.857813 6612 factory.go:656] Stopping watch factory\\\\nI1011 03:52:00.857815 6612 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1011 03:52:00.857835 6612 ovnkube.go:599] Stopped ovnkube\\\\nI1011 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:52:31Z\\\",\\\"message\\\":\\\"art network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:31Z is after 2025-08-24T17:21:41Z]\\\\nI1011 03:52:31.896403 6969 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-6xvgx\\\\nI1011 03:52:31.896411 6969 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-6xvgx\\\\nI1011 03:52:31.896420 6969 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-6xvgx in node crc\\\\nI1011 03:52:31.896427 6969 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-6xvgx after 0 failed attempt(s)\\\\nI1011 03:52:31.896434 6969 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-6xvgx\\\\nI1011 03:52:31.8\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:52:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.500169 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p24x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22c953c-0315-406e-9c5d-f1759837750e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p24x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.521010 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.539398 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.555504 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.573905 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83b1e937441999be327ba6a58c849e9c04956f5cea84da9f8f4aa476e1e79c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:52:21Z\\\",\\\"message\\\":\\\"2025-10-11T03:51:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_74aecb9d-f601-43f8-b48e-632d0155dabc\\\\n2025-10-11T03:51:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_74aecb9d-f601-43f8-b48e-632d0155dabc to /host/opt/cni/bin/\\\\n2025-10-11T03:51:36Z [verbose] multus-daemon started\\\\n2025-10-11T03:51:36Z [verbose] Readiness Indicator file check\\\\n2025-10-11T03:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.582353 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.582405 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.582424 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.582451 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.582470 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:32Z","lastTransitionTime":"2025-10-11T03:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.593535 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.609159 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.624784 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.644427 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.664243 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://257b2303215c18aa4500478fbfa39f9e1b95678c6d39ea51e4324c58f51dbd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.677636 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad406c24472c9ef7da82485a0600cf54f9002b83c0770c2ac86eca0f8779a870\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.685183 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.685230 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.685284 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.685311 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.685328 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:32Z","lastTransitionTime":"2025-10-11T03:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.691559 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"844146aa-36ac-4bee-8015-0d9850d94c4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cafb41d365e674a55638718df4e540ff77b6bd8e0d2d94bdb480b6617c0942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8f8a8bafbf88541375edb7c1b2134dec31a0d5cd46462a3271d24131bed4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hs9vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.788616 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.788691 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.788710 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.788759 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.788777 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:32Z","lastTransitionTime":"2025-10-11T03:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.834474 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b09975d-4c55-4b1a-b034-851ebf64cd3f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b90baa6dbb24ee4c2deea587bada1820917fd963caaf91210c25409b254c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fef6a4dbe93461c6352741acfc9439a22668f46dd725fcf80fb37d93c88e9d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f23f4ffaeac5c8606def100262484a45cd86296beb91735215c57955e45067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df874a11e199cea6058abe0e5bb63faf7246199fa02093f05720bc41dbb7a81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df874a11e199cea6058abe0e5bb63faf7246199fa02093f05720bc41dbb7a81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.866186 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3adee99370856f7d1dd95c3fed98740090f99e3d25c449f6f36170f0aeacd3e9\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:52:00Z\\\",\\\"message\\\":\\\":311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:52:00.856605 6612 reflector.go:311] Stopping reflector *v1.EndpointSlice (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.856659 6612 reflector.go:311] Stopping reflector *v1.NetworkPolicy (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.856768 6612 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI1011 03:52:00.857120 6612 reflector.go:311] Stopping reflector *v1.EgressFirewall (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressfirewall/v1/apis/informers/externalversions/factory.go:140\\\\nI1011 03:52:00.857719 6612 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI1011 03:52:00.857743 6612 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI1011 03:52:00.857756 6612 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI1011 03:52:00.857789 6612 handler.go:208] Removed *v1.Pod event handler 3\\\\nI1011 03:52:00.857813 6612 factory.go:656] Stopping watch factory\\\\nI1011 03:52:00.857815 6612 handler.go:208] Removed *v1.Pod event handler 6\\\\nI1011 03:52:00.857835 6612 ovnkube.go:599] Stopped ovnkube\\\\nI1011 0\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:59Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:52:31Z\\\",\\\"message\\\":\\\"art network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:31Z is after 2025-08-24T17:21:41Z]\\\\nI1011 03:52:31.896403 6969 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-6xvgx\\\\nI1011 03:52:31.896411 6969 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-6xvgx\\\\nI1011 03:52:31.896420 6969 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-6xvgx in node crc\\\\nI1011 03:52:31.896427 6969 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-6xvgx after 0 failed attempt(s)\\\\nI1011 03:52:31.896434 6969 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-6xvgx\\\\nI1011 03:52:31.8\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:52:30Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.882067 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p24x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22c953c-0315-406e-9c5d-f1759837750e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p24x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.901651 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.901695 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.901712 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.901734 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.901752 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:32Z","lastTransitionTime":"2025-10-11T03:52:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.905875 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.941675 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.963397 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:32 crc kubenswrapper[4967]: I1011 03:52:32.987090 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83b1e937441999be327ba6a58c849e9c04956f5cea84da9f8f4aa476e1e79c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:52:21Z\\\",\\\"message\\\":\\\"2025-10-11T03:51:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_74aecb9d-f601-43f8-b48e-632d0155dabc\\\\n2025-10-11T03:51:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_74aecb9d-f601-43f8-b48e-632d0155dabc to /host/opt/cni/bin/\\\\n2025-10-11T03:51:36Z [verbose] multus-daemon started\\\\n2025-10-11T03:51:36Z [verbose] Readiness Indicator file check\\\\n2025-10-11T03:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:32Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.003504 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.003562 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.003580 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.003603 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.003620 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:33Z","lastTransitionTime":"2025-10-11T03:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.007611 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.018308 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.029618 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.044240 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.056483 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://257b2303215c18aa4500478fbfa39f9e1b95678c6d39ea51e4324c58f51dbd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.067325 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad406c24472c9ef7da82485a0600cf54f9002b83c0770c2ac86eca0f8779a870\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.079268 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"844146aa-36ac-4bee-8015-0d9850d94c4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cafb41d365e674a55638718df4e540ff77b6bd8e0d2d94bdb480b6617c0942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8f8a8bafbf88541375edb7c1b2134dec31a0d5cd46462a3271d24131bed4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hs9vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.091391 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.102450 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.109399 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.109452 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.109470 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.109495 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.109511 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:33Z","lastTransitionTime":"2025-10-11T03:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.116906 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.127327 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.212141 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.212217 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.212240 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.212264 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.212281 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:33Z","lastTransitionTime":"2025-10-11T03:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.295629 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lw6dx_270ac8da-b130-4f4f-bbf7-655229b64446/ovnkube-controller/3.log" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.300565 4967 scope.go:117] "RemoveContainer" containerID="2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d" Oct 11 03:52:33 crc kubenswrapper[4967]: E1011 03:52:33.300816 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lw6dx_openshift-ovn-kubernetes(270ac8da-b130-4f4f-bbf7-655229b64446)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.314048 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.314097 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.314107 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.314120 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.314131 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:33Z","lastTransitionTime":"2025-10-11T03:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.320245 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.337783 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.359775 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83b1e937441999be327ba6a58c849e9c04956f5cea84da9f8f4aa476e1e79c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:52:21Z\\\",\\\"message\\\":\\\"2025-10-11T03:51:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_74aecb9d-f601-43f8-b48e-632d0155dabc\\\\n2025-10-11T03:51:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_74aecb9d-f601-43f8-b48e-632d0155dabc to /host/opt/cni/bin/\\\\n2025-10-11T03:51:36Z [verbose] multus-daemon started\\\\n2025-10-11T03:51:36Z [verbose] Readiness Indicator file check\\\\n2025-10-11T03:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.386137 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.404837 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.417531 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.417588 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.417606 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.417629 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.417649 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:33Z","lastTransitionTime":"2025-10-11T03:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.428376 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.448403 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.470778 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.494946 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://257b2303215c18aa4500478fbfa39f9e1b95678c6d39ea51e4324c58f51dbd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.512637 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad406c24472c9ef7da82485a0600cf54f9002b83c0770c2ac86eca0f8779a870\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.520316 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.520396 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.520416 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.520444 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.520464 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:33Z","lastTransitionTime":"2025-10-11T03:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.533742 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"844146aa-36ac-4bee-8015-0d9850d94c4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cafb41d365e674a55638718df4e540ff77b6bd8e0d2d94bdb480b6617c0942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8f8a8bafbf88541375edb7c1b2134dec31a0d5cd46462a3271d24131bed4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hs9vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.557386 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.573718 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.590323 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.608938 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.622218 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b09975d-4c55-4b1a-b034-851ebf64cd3f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b90baa6dbb24ee4c2deea587bada1820917fd963caaf91210c25409b254c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fef6a4dbe93461c6352741acfc9439a22668f46dd725fcf80fb37d93c88e9d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f23f4ffaeac5c8606def100262484a45cd86296beb91735215c57955e45067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df874a11e199cea6058abe0e5bb63faf7246199fa02093f05720bc41dbb7a81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df874a11e199cea6058abe0e5bb63faf7246199fa02093f05720bc41dbb7a81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.623025 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.623139 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.623161 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.623186 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.623204 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:33Z","lastTransitionTime":"2025-10-11T03:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.651176 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:52:31Z\\\",\\\"message\\\":\\\"art network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:31Z is after 2025-08-24T17:21:41Z]\\\\nI1011 03:52:31.896403 6969 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-6xvgx\\\\nI1011 03:52:31.896411 6969 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-6xvgx\\\\nI1011 03:52:31.896420 6969 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-6xvgx in node crc\\\\nI1011 03:52:31.896427 6969 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-6xvgx after 0 failed attempt(s)\\\\nI1011 03:52:31.896434 6969 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-6xvgx\\\\nI1011 03:52:31.8\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:52:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lw6dx_openshift-ovn-kubernetes(270ac8da-b130-4f4f-bbf7-655229b64446)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.664757 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p24x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22c953c-0315-406e-9c5d-f1759837750e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p24x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:33Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.726657 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.726721 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.726744 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.726776 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.726795 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:33Z","lastTransitionTime":"2025-10-11T03:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.813852 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.813852 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.813899 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:33 crc kubenswrapper[4967]: E1011 03:52:33.814648 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:52:33 crc kubenswrapper[4967]: E1011 03:52:33.814358 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.814121 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:33 crc kubenswrapper[4967]: E1011 03:52:33.814758 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:52:33 crc kubenswrapper[4967]: E1011 03:52:33.814944 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.829710 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.829946 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.830646 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.830911 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.831153 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:33Z","lastTransitionTime":"2025-10-11T03:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.934179 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.934256 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.934280 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.934306 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:33 crc kubenswrapper[4967]: I1011 03:52:33.934323 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:33Z","lastTransitionTime":"2025-10-11T03:52:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.037348 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.037443 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.037462 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.037485 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.037502 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:34Z","lastTransitionTime":"2025-10-11T03:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.140983 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.141063 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.141110 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.141141 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.141158 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:34Z","lastTransitionTime":"2025-10-11T03:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.244195 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.244265 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.244282 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.244303 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.244317 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:34Z","lastTransitionTime":"2025-10-11T03:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.346849 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.346892 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.346904 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.346922 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.346935 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:34Z","lastTransitionTime":"2025-10-11T03:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.449348 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.449439 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.449450 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.449468 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.449479 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:34Z","lastTransitionTime":"2025-10-11T03:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.552152 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.552215 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.552233 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.552257 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.552280 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:34Z","lastTransitionTime":"2025-10-11T03:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.655628 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.655712 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.655730 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.655754 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.655777 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:34Z","lastTransitionTime":"2025-10-11T03:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.759056 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.759197 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.759442 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.759505 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.759525 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:34Z","lastTransitionTime":"2025-10-11T03:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.832585 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.861907 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.861977 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.862001 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.862028 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.862051 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:34Z","lastTransitionTime":"2025-10-11T03:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.964964 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.965020 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.965036 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.965057 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:34 crc kubenswrapper[4967]: I1011 03:52:34.965101 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:34Z","lastTransitionTime":"2025-10-11T03:52:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.068332 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.068460 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.068495 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.068523 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.068561 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:35Z","lastTransitionTime":"2025-10-11T03:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.171851 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.171932 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.171960 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.171988 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.172004 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:35Z","lastTransitionTime":"2025-10-11T03:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.275523 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.275593 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.275609 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.275633 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.275651 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:35Z","lastTransitionTime":"2025-10-11T03:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.379043 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.379129 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.379146 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.379171 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.379191 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:35Z","lastTransitionTime":"2025-10-11T03:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.482360 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.482425 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.482443 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.482466 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.482484 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:35Z","lastTransitionTime":"2025-10-11T03:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.585659 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.585698 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.585714 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.585739 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.585757 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:35Z","lastTransitionTime":"2025-10-11T03:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.637484 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.637584 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:35 crc kubenswrapper[4967]: E1011 03:52:35.637631 4967 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Oct 11 03:52:35 crc kubenswrapper[4967]: E1011 03:52:35.637736 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-11 03:53:39.637708657 +0000 UTC m=+147.600917620 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Oct 11 03:52:35 crc kubenswrapper[4967]: E1011 03:52:35.637779 4967 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 11 03:52:35 crc kubenswrapper[4967]: E1011 03:52:35.637859 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2025-10-11 03:53:39.63783584 +0000 UTC m=+147.601044803 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.688368 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.688426 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.688442 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.688465 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.688483 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:35Z","lastTransitionTime":"2025-10-11T03:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.739140 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.739270 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:52:35 crc kubenswrapper[4967]: E1011 03:52:35.739468 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 11 03:52:35 crc kubenswrapper[4967]: E1011 03:52:35.739497 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 11 03:52:35 crc kubenswrapper[4967]: E1011 03:52:35.739500 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Oct 11 03:52:35 crc kubenswrapper[4967]: E1011 03:52:35.739583 4967 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Oct 11 03:52:35 crc kubenswrapper[4967]: E1011 03:52:35.739518 4967 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:52:35 crc kubenswrapper[4967]: E1011 03:52:35.739613 4967 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:52:35 crc kubenswrapper[4967]: E1011 03:52:35.739696 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2025-10-11 03:53:39.739671787 +0000 UTC m=+147.702880760 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:52:35 crc kubenswrapper[4967]: E1011 03:52:35.739725 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2025-10-11 03:53:39.739712908 +0000 UTC m=+147.702921881 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.790662 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.790739 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.790756 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.790782 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.790801 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:35Z","lastTransitionTime":"2025-10-11T03:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.813792 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.813800 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:35 crc kubenswrapper[4967]: E1011 03:52:35.814005 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.813809 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.813800 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:52:35 crc kubenswrapper[4967]: E1011 03:52:35.814110 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:52:35 crc kubenswrapper[4967]: E1011 03:52:35.814239 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:52:35 crc kubenswrapper[4967]: E1011 03:52:35.814390 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.840625 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:52:35 crc kubenswrapper[4967]: E1011 03:52:35.840863 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:39.840835675 +0000 UTC m=+147.804044618 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.894251 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.894290 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.894307 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.894329 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.894348 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:35Z","lastTransitionTime":"2025-10-11T03:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.997110 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.997166 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.997183 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.997205 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:35 crc kubenswrapper[4967]: I1011 03:52:35.997222 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:35Z","lastTransitionTime":"2025-10-11T03:52:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.100009 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.100113 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.100140 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.100169 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.100192 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:36Z","lastTransitionTime":"2025-10-11T03:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.203173 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.203344 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.203417 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.203454 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.203476 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:36Z","lastTransitionTime":"2025-10-11T03:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.306722 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.306797 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.306823 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.306853 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.306880 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:36Z","lastTransitionTime":"2025-10-11T03:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.410562 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.410612 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.410642 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.410669 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.410692 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:36Z","lastTransitionTime":"2025-10-11T03:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.514504 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.514622 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.514646 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.514678 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.514699 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:36Z","lastTransitionTime":"2025-10-11T03:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.617817 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.617875 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.617891 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.617917 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.617936 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:36Z","lastTransitionTime":"2025-10-11T03:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.721425 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.721483 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.721498 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.721521 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.721538 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:36Z","lastTransitionTime":"2025-10-11T03:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.823775 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.823853 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.823871 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.823900 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.823918 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:36Z","lastTransitionTime":"2025-10-11T03:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.927141 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.927232 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.927264 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.927295 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:36 crc kubenswrapper[4967]: I1011 03:52:36.927312 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:36Z","lastTransitionTime":"2025-10-11T03:52:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.030997 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.031109 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.031130 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.031155 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.031180 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:37Z","lastTransitionTime":"2025-10-11T03:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.134745 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.134800 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.134815 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.134838 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.134853 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:37Z","lastTransitionTime":"2025-10-11T03:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.237589 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.237651 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.237668 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.237692 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.237710 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:37Z","lastTransitionTime":"2025-10-11T03:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.340727 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.340814 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.340843 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.340878 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.340904 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:37Z","lastTransitionTime":"2025-10-11T03:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.444463 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.444892 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.444910 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.444934 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.444951 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:37Z","lastTransitionTime":"2025-10-11T03:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.548258 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.548335 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.548359 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.548382 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.548400 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:37Z","lastTransitionTime":"2025-10-11T03:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.654524 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.655278 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.655315 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.655348 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.655370 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:37Z","lastTransitionTime":"2025-10-11T03:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.758480 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.758547 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.758566 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.758591 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.758611 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:37Z","lastTransitionTime":"2025-10-11T03:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.814235 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.814251 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.814325 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.814381 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:37 crc kubenswrapper[4967]: E1011 03:52:37.814534 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:52:37 crc kubenswrapper[4967]: E1011 03:52:37.814725 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:52:37 crc kubenswrapper[4967]: E1011 03:52:37.814768 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:52:37 crc kubenswrapper[4967]: E1011 03:52:37.814833 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.861767 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.861847 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.861864 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.861886 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.861901 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:37Z","lastTransitionTime":"2025-10-11T03:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.964649 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.964695 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.964711 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.964734 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:37 crc kubenswrapper[4967]: I1011 03:52:37.964751 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:37Z","lastTransitionTime":"2025-10-11T03:52:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.067840 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.067920 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.067937 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.067961 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.067977 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:38Z","lastTransitionTime":"2025-10-11T03:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.170444 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.170502 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.170520 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.170544 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.170563 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:38Z","lastTransitionTime":"2025-10-11T03:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.274017 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.274361 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.274561 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.274691 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.274829 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:38Z","lastTransitionTime":"2025-10-11T03:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.378785 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.379247 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.379443 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.379659 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.380228 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:38Z","lastTransitionTime":"2025-10-11T03:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.483647 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.483682 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.483692 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.483708 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.483718 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:38Z","lastTransitionTime":"2025-10-11T03:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.586344 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.586383 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.586392 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.586408 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.586419 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:38Z","lastTransitionTime":"2025-10-11T03:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.689145 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.689183 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.689196 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.689212 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.689223 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:38Z","lastTransitionTime":"2025-10-11T03:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.792323 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.792391 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.792414 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.792445 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.792467 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:38Z","lastTransitionTime":"2025-10-11T03:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.895681 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.896013 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.896237 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.896495 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:38 crc kubenswrapper[4967]: I1011 03:52:38.896734 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:38Z","lastTransitionTime":"2025-10-11T03:52:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.000522 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.000584 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.000606 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.000631 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.000651 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:39Z","lastTransitionTime":"2025-10-11T03:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.103025 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.103056 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.103068 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.103106 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.103118 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:39Z","lastTransitionTime":"2025-10-11T03:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.207819 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.207894 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.207914 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.207943 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.207961 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:39Z","lastTransitionTime":"2025-10-11T03:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.310944 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.311015 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.311026 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.311097 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.311113 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:39Z","lastTransitionTime":"2025-10-11T03:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.414512 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.414578 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.414595 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.414620 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.414639 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:39Z","lastTransitionTime":"2025-10-11T03:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.518266 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.518358 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.518382 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.518413 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.518440 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:39Z","lastTransitionTime":"2025-10-11T03:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.621775 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.621860 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.621881 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.621915 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.621938 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:39Z","lastTransitionTime":"2025-10-11T03:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.724747 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.724799 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.724815 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.724837 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.724852 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:39Z","lastTransitionTime":"2025-10-11T03:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.814248 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.814309 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:52:39 crc kubenswrapper[4967]: E1011 03:52:39.814426 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.814496 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.814550 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:39 crc kubenswrapper[4967]: E1011 03:52:39.814621 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:52:39 crc kubenswrapper[4967]: E1011 03:52:39.815190 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:52:39 crc kubenswrapper[4967]: E1011 03:52:39.815180 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.827867 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.827923 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.827943 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.827969 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.827987 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:39Z","lastTransitionTime":"2025-10-11T03:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.931294 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.931348 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.931362 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.931385 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:39 crc kubenswrapper[4967]: I1011 03:52:39.931400 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:39Z","lastTransitionTime":"2025-10-11T03:52:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.014352 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.014421 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.014436 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.014458 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.014877 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:40Z","lastTransitionTime":"2025-10-11T03:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:40 crc kubenswrapper[4967]: E1011 03:52:40.030714 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:40Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.034735 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.034816 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.034831 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.034853 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.034867 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:40Z","lastTransitionTime":"2025-10-11T03:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:40 crc kubenswrapper[4967]: E1011 03:52:40.055519 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:40Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.061716 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.061771 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.061784 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.061803 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.061815 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:40Z","lastTransitionTime":"2025-10-11T03:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:40 crc kubenswrapper[4967]: E1011 03:52:40.081349 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:40Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.086119 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.086217 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.086235 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.086255 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.086304 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:40Z","lastTransitionTime":"2025-10-11T03:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:40 crc kubenswrapper[4967]: E1011 03:52:40.105421 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:40Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.110356 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.110415 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.110427 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.110447 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.110460 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:40Z","lastTransitionTime":"2025-10-11T03:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:40 crc kubenswrapper[4967]: E1011 03:52:40.128970 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:40Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:40Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:40 crc kubenswrapper[4967]: E1011 03:52:40.129470 4967 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.131422 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.131538 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.131640 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.131729 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.131821 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:40Z","lastTransitionTime":"2025-10-11T03:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.233836 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.233899 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.233916 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.233943 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.233962 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:40Z","lastTransitionTime":"2025-10-11T03:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.336448 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.336487 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.336499 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.336515 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.336527 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:40Z","lastTransitionTime":"2025-10-11T03:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.439483 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.439529 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.439545 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.439568 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.439585 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:40Z","lastTransitionTime":"2025-10-11T03:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.541939 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.541998 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.542015 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.542039 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.542057 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:40Z","lastTransitionTime":"2025-10-11T03:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.645010 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.645052 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.645069 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.645128 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.645145 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:40Z","lastTransitionTime":"2025-10-11T03:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.747773 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.747833 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.747849 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.747876 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.747893 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:40Z","lastTransitionTime":"2025-10-11T03:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.850673 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.850717 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.850731 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.850747 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.850758 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:40Z","lastTransitionTime":"2025-10-11T03:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.954029 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.954109 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.954129 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.954155 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:40 crc kubenswrapper[4967]: I1011 03:52:40.954173 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:40Z","lastTransitionTime":"2025-10-11T03:52:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.057485 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.057525 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.057539 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.057558 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.057571 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:41Z","lastTransitionTime":"2025-10-11T03:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.160201 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.160261 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.160284 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.160316 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.160343 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:41Z","lastTransitionTime":"2025-10-11T03:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.262946 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.262988 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.263002 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.263019 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.263031 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:41Z","lastTransitionTime":"2025-10-11T03:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.365825 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.365890 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.365910 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.365935 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.365953 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:41Z","lastTransitionTime":"2025-10-11T03:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.469548 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.469611 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.469628 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.469652 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.469671 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:41Z","lastTransitionTime":"2025-10-11T03:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.572047 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.572159 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.572177 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.572202 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.572220 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:41Z","lastTransitionTime":"2025-10-11T03:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.675620 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.675684 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.675702 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.675731 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.675748 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:41Z","lastTransitionTime":"2025-10-11T03:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.778774 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.778835 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.778852 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.778875 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.778893 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:41Z","lastTransitionTime":"2025-10-11T03:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.814404 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.814445 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.814477 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.814407 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:52:41 crc kubenswrapper[4967]: E1011 03:52:41.814621 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:52:41 crc kubenswrapper[4967]: E1011 03:52:41.814741 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:52:41 crc kubenswrapper[4967]: E1011 03:52:41.814914 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:52:41 crc kubenswrapper[4967]: E1011 03:52:41.815039 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.882151 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.882213 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.882231 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.882257 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.882275 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:41Z","lastTransitionTime":"2025-10-11T03:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.985381 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.985446 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.985464 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.985491 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:41 crc kubenswrapper[4967]: I1011 03:52:41.985510 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:41Z","lastTransitionTime":"2025-10-11T03:52:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.088498 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.088572 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.088591 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.088615 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.088633 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:42Z","lastTransitionTime":"2025-10-11T03:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.191575 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.191639 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.191657 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.191684 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.191706 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:42Z","lastTransitionTime":"2025-10-11T03:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.299311 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.299411 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.299503 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.299579 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.299600 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:42Z","lastTransitionTime":"2025-10-11T03:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.402342 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.402403 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.402420 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.402444 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.402461 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:42Z","lastTransitionTime":"2025-10-11T03:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.505465 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.505526 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.505543 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.505567 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.505584 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:42Z","lastTransitionTime":"2025-10-11T03:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.608912 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.608966 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.608983 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.609005 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.609023 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:42Z","lastTransitionTime":"2025-10-11T03:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.712006 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.712107 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.712129 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.712150 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.712166 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:42Z","lastTransitionTime":"2025-10-11T03:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.814902 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.814959 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.814977 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.815001 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.815019 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:42Z","lastTransitionTime":"2025-10-11T03:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.833190 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47e69eed-ac05-4d04-a290-580433d14207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fc96d2d9f48b01b4c4f1f4bc3e058427ffc11b3b73c0ad8a89dbd690a098b67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e84c5697ab006ad26c4ebc739bf09997d679c605f84ffe355d30f2b32f4f0e54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e84c5697ab006ad26c4ebc739bf09997d679c605f84ffe355d30f2b32f4f0e54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.855554 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.875970 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.897412 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.915617 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.917331 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.917357 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.917366 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.917380 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.917390 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:42Z","lastTransitionTime":"2025-10-11T03:52:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.935059 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b09975d-4c55-4b1a-b034-851ebf64cd3f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b90baa6dbb24ee4c2deea587bada1820917fd963caaf91210c25409b254c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fef6a4dbe93461c6352741acfc9439a22668f46dd725fcf80fb37d93c88e9d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f23f4ffaeac5c8606def100262484a45cd86296beb91735215c57955e45067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df874a11e199cea6058abe0e5bb63faf7246199fa02093f05720bc41dbb7a81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df874a11e199cea6058abe0e5bb63faf7246199fa02093f05720bc41dbb7a81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.968569 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:52:31Z\\\",\\\"message\\\":\\\"art network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:31Z is after 2025-08-24T17:21:41Z]\\\\nI1011 03:52:31.896403 6969 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-6xvgx\\\\nI1011 03:52:31.896411 6969 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-6xvgx\\\\nI1011 03:52:31.896420 6969 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-6xvgx in node crc\\\\nI1011 03:52:31.896427 6969 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-6xvgx after 0 failed attempt(s)\\\\nI1011 03:52:31.896434 6969 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-6xvgx\\\\nI1011 03:52:31.8\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:52:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lw6dx_openshift-ovn-kubernetes(270ac8da-b130-4f4f-bbf7-655229b64446)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:42 crc kubenswrapper[4967]: I1011 03:52:42.989432 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p24x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22c953c-0315-406e-9c5d-f1759837750e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p24x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:42Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.009328 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.020774 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.020830 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.020848 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.020873 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.020891 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:43Z","lastTransitionTime":"2025-10-11T03:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.027474 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.049041 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83b1e937441999be327ba6a58c849e9c04956f5cea84da9f8f4aa476e1e79c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:52:21Z\\\",\\\"message\\\":\\\"2025-10-11T03:51:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_74aecb9d-f601-43f8-b48e-632d0155dabc\\\\n2025-10-11T03:51:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_74aecb9d-f601-43f8-b48e-632d0155dabc to /host/opt/cni/bin/\\\\n2025-10-11T03:51:36Z [verbose] multus-daemon started\\\\n2025-10-11T03:51:36Z [verbose] Readiness Indicator file check\\\\n2025-10-11T03:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.083788 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.101379 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.122335 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.123679 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.123740 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.123763 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.123926 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.123992 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:43Z","lastTransitionTime":"2025-10-11T03:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.140291 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.161459 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.184424 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://257b2303215c18aa4500478fbfa39f9e1b95678c6d39ea51e4324c58f51dbd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.201045 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad406c24472c9ef7da82485a0600cf54f9002b83c0770c2ac86eca0f8779a870\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.218826 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"844146aa-36ac-4bee-8015-0d9850d94c4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cafb41d365e674a55638718df4e540ff77b6bd8e0d2d94bdb480b6617c0942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8f8a8bafbf88541375edb7c1b2134dec31a0d5cd46462a3271d24131bed4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hs9vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:43Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.227578 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.227628 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.227646 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.227673 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.227693 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:43Z","lastTransitionTime":"2025-10-11T03:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.330697 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.330746 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.330762 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.330781 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.330801 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:43Z","lastTransitionTime":"2025-10-11T03:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.433647 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.433709 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.433724 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.433755 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.433771 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:43Z","lastTransitionTime":"2025-10-11T03:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.536505 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.536582 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.536598 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.536624 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.536647 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:43Z","lastTransitionTime":"2025-10-11T03:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.639309 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.639388 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.639410 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.639437 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.639454 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:43Z","lastTransitionTime":"2025-10-11T03:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.741994 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.742128 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.742150 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.742179 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.742200 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:43Z","lastTransitionTime":"2025-10-11T03:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.814864 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.814903 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.814883 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.815484 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:52:43 crc kubenswrapper[4967]: E1011 03:52:43.815689 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:52:43 crc kubenswrapper[4967]: E1011 03:52:43.815807 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.816051 4967 scope.go:117] "RemoveContainer" containerID="2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d" Oct 11 03:52:43 crc kubenswrapper[4967]: E1011 03:52:43.816228 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:52:43 crc kubenswrapper[4967]: E1011 03:52:43.816362 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lw6dx_openshift-ovn-kubernetes(270ac8da-b130-4f4f-bbf7-655229b64446)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" Oct 11 03:52:43 crc kubenswrapper[4967]: E1011 03:52:43.816365 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.845201 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.845273 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.845287 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.845309 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.845321 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:43Z","lastTransitionTime":"2025-10-11T03:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.947998 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.948063 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.948115 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.948140 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:43 crc kubenswrapper[4967]: I1011 03:52:43.948156 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:43Z","lastTransitionTime":"2025-10-11T03:52:43Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.052286 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.052398 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.052416 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.052450 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.052478 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:44Z","lastTransitionTime":"2025-10-11T03:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.156160 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.156219 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.156237 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.156262 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.156281 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:44Z","lastTransitionTime":"2025-10-11T03:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.260067 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.260427 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.260557 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.260686 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.260819 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:44Z","lastTransitionTime":"2025-10-11T03:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.363477 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.363540 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.363559 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.363584 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.363603 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:44Z","lastTransitionTime":"2025-10-11T03:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.471991 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.472046 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.472059 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.472098 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.472112 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:44Z","lastTransitionTime":"2025-10-11T03:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.575894 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.575968 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.575978 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.575999 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.576013 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:44Z","lastTransitionTime":"2025-10-11T03:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.679704 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.679790 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.679804 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.679828 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.679845 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:44Z","lastTransitionTime":"2025-10-11T03:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.783670 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.783768 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.783787 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.783842 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.783862 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:44Z","lastTransitionTime":"2025-10-11T03:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.887004 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.887116 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.887140 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.887169 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.887188 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:44Z","lastTransitionTime":"2025-10-11T03:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.990604 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.990675 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.990696 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.990726 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:44 crc kubenswrapper[4967]: I1011 03:52:44.990749 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:44Z","lastTransitionTime":"2025-10-11T03:52:44Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.094045 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.094135 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.094152 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.094175 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.094192 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:45Z","lastTransitionTime":"2025-10-11T03:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.196389 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.196446 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.196466 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.196494 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.196516 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:45Z","lastTransitionTime":"2025-10-11T03:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.299314 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.299384 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.299407 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.299438 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.299462 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:45Z","lastTransitionTime":"2025-10-11T03:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.402412 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.402493 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.402518 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.402548 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.402570 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:45Z","lastTransitionTime":"2025-10-11T03:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.505734 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.505790 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.505806 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.505827 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.505843 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:45Z","lastTransitionTime":"2025-10-11T03:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.607766 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.607822 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.607844 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.607869 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.607892 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:45Z","lastTransitionTime":"2025-10-11T03:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.711234 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.711293 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.711319 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.711343 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.711363 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:45Z","lastTransitionTime":"2025-10-11T03:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.813959 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.813997 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.814000 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:52:45 crc kubenswrapper[4967]: E1011 03:52:45.814201 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.814259 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:45 crc kubenswrapper[4967]: E1011 03:52:45.814385 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:52:45 crc kubenswrapper[4967]: E1011 03:52:45.814518 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:52:45 crc kubenswrapper[4967]: E1011 03:52:45.814710 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.815121 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.815169 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.815191 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.815217 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.815237 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:45Z","lastTransitionTime":"2025-10-11T03:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.918351 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.918407 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.918423 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.918447 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:45 crc kubenswrapper[4967]: I1011 03:52:45.918464 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:45Z","lastTransitionTime":"2025-10-11T03:52:45Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.020903 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.020959 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.020974 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.021002 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.021021 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:46Z","lastTransitionTime":"2025-10-11T03:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.124466 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.124534 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.124551 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.124571 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.124592 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:46Z","lastTransitionTime":"2025-10-11T03:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.227216 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.227263 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.227279 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.227302 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.227320 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:46Z","lastTransitionTime":"2025-10-11T03:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.329863 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.329923 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.329940 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.329969 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.329986 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:46Z","lastTransitionTime":"2025-10-11T03:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.432056 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.432158 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.432180 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.432211 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.432238 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:46Z","lastTransitionTime":"2025-10-11T03:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.535592 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.535657 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.535679 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.535708 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.535730 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:46Z","lastTransitionTime":"2025-10-11T03:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.638257 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.638316 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.638337 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.638366 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.638390 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:46Z","lastTransitionTime":"2025-10-11T03:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.740447 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.740513 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.740532 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.740557 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.740575 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:46Z","lastTransitionTime":"2025-10-11T03:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.843787 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.843852 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.843874 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.843903 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.843923 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:46Z","lastTransitionTime":"2025-10-11T03:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.946588 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.946647 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.946666 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.946689 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:46 crc kubenswrapper[4967]: I1011 03:52:46.946706 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:46Z","lastTransitionTime":"2025-10-11T03:52:46Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.049051 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.049177 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.049203 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.049234 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.049295 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:47Z","lastTransitionTime":"2025-10-11T03:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.151829 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.151889 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.151911 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.151937 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.151959 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:47Z","lastTransitionTime":"2025-10-11T03:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.254655 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.254704 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.254718 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.254738 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.254749 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:47Z","lastTransitionTime":"2025-10-11T03:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.357640 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.357677 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.357688 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.357705 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.357717 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:47Z","lastTransitionTime":"2025-10-11T03:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.460827 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.460869 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.460880 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.460897 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.460910 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:47Z","lastTransitionTime":"2025-10-11T03:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.563420 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.563493 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.563515 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.563544 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.563568 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:47Z","lastTransitionTime":"2025-10-11T03:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.666484 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.666522 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.666529 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.666546 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.666558 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:47Z","lastTransitionTime":"2025-10-11T03:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.770359 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.770397 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.770408 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.770423 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.770435 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:47Z","lastTransitionTime":"2025-10-11T03:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.814159 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.814335 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.814399 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.814426 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:47 crc kubenswrapper[4967]: E1011 03:52:47.814536 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:52:47 crc kubenswrapper[4967]: E1011 03:52:47.814636 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:52:47 crc kubenswrapper[4967]: E1011 03:52:47.814729 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:52:47 crc kubenswrapper[4967]: E1011 03:52:47.814792 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.872764 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.872829 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.872851 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.872880 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.872902 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:47Z","lastTransitionTime":"2025-10-11T03:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.975197 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.975237 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.975248 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.975266 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:47 crc kubenswrapper[4967]: I1011 03:52:47.975276 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:47Z","lastTransitionTime":"2025-10-11T03:52:47Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.077584 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.077635 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.077652 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.077677 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.077694 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:48Z","lastTransitionTime":"2025-10-11T03:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.180184 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.180229 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.180241 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.180283 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.180299 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:48Z","lastTransitionTime":"2025-10-11T03:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.283603 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.283657 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.283675 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.283704 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.283724 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:48Z","lastTransitionTime":"2025-10-11T03:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.386098 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.386152 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.386168 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.386190 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.386206 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:48Z","lastTransitionTime":"2025-10-11T03:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.489045 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.489098 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.489109 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.489125 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.489137 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:48Z","lastTransitionTime":"2025-10-11T03:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.591757 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.591789 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.591801 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.591816 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.591826 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:48Z","lastTransitionTime":"2025-10-11T03:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.694857 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.694922 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.694947 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.694978 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.695002 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:48Z","lastTransitionTime":"2025-10-11T03:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.798185 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.798263 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.798282 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.798759 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.798827 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:48Z","lastTransitionTime":"2025-10-11T03:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.901999 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.902058 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.902101 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.902130 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:48 crc kubenswrapper[4967]: I1011 03:52:48.902149 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:48Z","lastTransitionTime":"2025-10-11T03:52:48Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.005349 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.005426 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.005447 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.005472 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.005489 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:49Z","lastTransitionTime":"2025-10-11T03:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.108618 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.108675 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.108693 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.108727 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.108754 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:49Z","lastTransitionTime":"2025-10-11T03:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.212161 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.212187 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.212195 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.212207 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.212214 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:49Z","lastTransitionTime":"2025-10-11T03:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.315127 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.315173 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.315184 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.315207 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.315218 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:49Z","lastTransitionTime":"2025-10-11T03:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.418510 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.418559 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.418575 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.418599 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.418615 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:49Z","lastTransitionTime":"2025-10-11T03:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.521380 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.521429 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.521448 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.521472 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.521490 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:49Z","lastTransitionTime":"2025-10-11T03:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.625042 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.625133 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.625152 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.625179 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.625197 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:49Z","lastTransitionTime":"2025-10-11T03:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.729038 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.729160 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.729188 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.729218 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.729239 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:49Z","lastTransitionTime":"2025-10-11T03:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.814499 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.814620 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:52:49 crc kubenswrapper[4967]: E1011 03:52:49.814820 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.814870 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.814940 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:52:49 crc kubenswrapper[4967]: E1011 03:52:49.815323 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:52:49 crc kubenswrapper[4967]: E1011 03:52:49.815479 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:52:49 crc kubenswrapper[4967]: E1011 03:52:49.815619 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.831657 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.831711 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.831729 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.831753 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.831773 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:49Z","lastTransitionTime":"2025-10-11T03:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.934302 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.934344 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.934355 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.934370 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:49 crc kubenswrapper[4967]: I1011 03:52:49.934381 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:49Z","lastTransitionTime":"2025-10-11T03:52:49Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.037803 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.037858 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.037881 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.037904 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.037921 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:50Z","lastTransitionTime":"2025-10-11T03:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.141301 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.141370 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.141389 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.141415 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.141434 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:50Z","lastTransitionTime":"2025-10-11T03:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.244267 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.244334 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.244357 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.244387 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.244412 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:50Z","lastTransitionTime":"2025-10-11T03:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.281051 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.281146 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.281164 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.281190 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.281212 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:50Z","lastTransitionTime":"2025-10-11T03:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:50 crc kubenswrapper[4967]: E1011 03:52:50.302030 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:50Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.306838 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.306910 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.306933 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.306962 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.306983 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:50Z","lastTransitionTime":"2025-10-11T03:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:50 crc kubenswrapper[4967]: E1011 03:52:50.325760 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:50Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.330880 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.330939 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.330951 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.330970 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.330986 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:50Z","lastTransitionTime":"2025-10-11T03:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:50 crc kubenswrapper[4967]: E1011 03:52:50.350423 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:50Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.355812 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.355887 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.355910 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.355945 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.355966 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:50Z","lastTransitionTime":"2025-10-11T03:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:50 crc kubenswrapper[4967]: E1011 03:52:50.377889 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:50Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.382568 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.382618 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.382640 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.382661 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.382677 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:50Z","lastTransitionTime":"2025-10-11T03:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:50 crc kubenswrapper[4967]: E1011 03:52:50.402874 4967 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404560Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865360Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:50Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"2f687e2b-93fa-4909-96cb-dd8768ce3482\\\",\\\"systemUUID\\\":\\\"39750116-adfb-459d-8c17-e38159f0ae3a\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:50Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:50 crc kubenswrapper[4967]: E1011 03:52:50.403329 4967 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.405759 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.405809 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.405827 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.405850 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.405869 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:50Z","lastTransitionTime":"2025-10-11T03:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.508677 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.508741 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.508764 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.508792 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.508891 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:50Z","lastTransitionTime":"2025-10-11T03:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.611932 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.612021 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.612044 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.612112 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.612137 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:50Z","lastTransitionTime":"2025-10-11T03:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.715357 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.715406 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.715418 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.715434 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.715789 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:50Z","lastTransitionTime":"2025-10-11T03:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.818698 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.818738 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.818748 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.818761 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.818773 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:50Z","lastTransitionTime":"2025-10-11T03:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.921487 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.921546 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.921563 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.921587 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:50 crc kubenswrapper[4967]: I1011 03:52:50.921606 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:50Z","lastTransitionTime":"2025-10-11T03:52:50Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.025193 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.025276 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.025299 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.025330 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.025351 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:51Z","lastTransitionTime":"2025-10-11T03:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.128880 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.128945 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.128967 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.128997 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.129018 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:51Z","lastTransitionTime":"2025-10-11T03:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.231467 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.231532 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.231553 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.231581 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.231603 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:51Z","lastTransitionTime":"2025-10-11T03:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.334498 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.334549 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.334562 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.334581 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.334596 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:51Z","lastTransitionTime":"2025-10-11T03:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.438302 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.438393 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.438418 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.438452 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.438475 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:51Z","lastTransitionTime":"2025-10-11T03:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.541158 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.541223 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.541242 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.541266 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.541286 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:51Z","lastTransitionTime":"2025-10-11T03:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.643852 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.643903 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.643921 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.643942 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.643958 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:51Z","lastTransitionTime":"2025-10-11T03:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.746692 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.746777 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.746800 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.746828 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.746850 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:51Z","lastTransitionTime":"2025-10-11T03:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.814846 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.814967 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.815025 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.815160 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:51 crc kubenswrapper[4967]: E1011 03:52:51.815257 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:52:51 crc kubenswrapper[4967]: E1011 03:52:51.815312 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:52:51 crc kubenswrapper[4967]: E1011 03:52:51.815461 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:52:51 crc kubenswrapper[4967]: E1011 03:52:51.815619 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.849397 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.849472 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.849496 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.849527 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.849545 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:51Z","lastTransitionTime":"2025-10-11T03:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.953633 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.953703 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.953730 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.953764 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:51 crc kubenswrapper[4967]: I1011 03:52:51.953789 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:51Z","lastTransitionTime":"2025-10-11T03:52:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.021163 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d22c953c-0315-406e-9c5d-f1759837750e-metrics-certs\") pod \"network-metrics-daemon-p24x8\" (UID: \"d22c953c-0315-406e-9c5d-f1759837750e\") " pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:52 crc kubenswrapper[4967]: E1011 03:52:52.021610 4967 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Oct 11 03:52:52 crc kubenswrapper[4967]: E1011 03:52:52.021746 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/d22c953c-0315-406e-9c5d-f1759837750e-metrics-certs podName:d22c953c-0315-406e-9c5d-f1759837750e nodeName:}" failed. No retries permitted until 2025-10-11 03:53:56.021715851 +0000 UTC m=+163.984924824 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/d22c953c-0315-406e-9c5d-f1759837750e-metrics-certs") pod "network-metrics-daemon-p24x8" (UID: "d22c953c-0315-406e-9c5d-f1759837750e") : object "openshift-multus"/"metrics-daemon-secret" not registered Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.058149 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.058223 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.058240 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.058270 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.058288 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:52Z","lastTransitionTime":"2025-10-11T03:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.162377 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.162453 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.162471 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.162496 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.162515 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:52Z","lastTransitionTime":"2025-10-11T03:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.265717 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.265776 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.265792 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.265819 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.265838 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:52Z","lastTransitionTime":"2025-10-11T03:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.368378 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.368435 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.368452 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.368475 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.368492 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:52Z","lastTransitionTime":"2025-10-11T03:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.471861 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.471916 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.471932 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.471953 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.471970 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:52Z","lastTransitionTime":"2025-10-11T03:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.574474 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.574518 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.574536 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.574556 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.574571 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:52Z","lastTransitionTime":"2025-10-11T03:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.678097 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.678176 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.678194 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.678221 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.678240 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:52Z","lastTransitionTime":"2025-10-11T03:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.781671 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.781734 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.781751 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.781777 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.781798 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:52Z","lastTransitionTime":"2025-10-11T03:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.839143 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7b09975d-4c55-4b1a-b034-851ebf64cd3f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b2b90baa6dbb24ee4c2deea587bada1820917fd963caaf91210c25409b254c60\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fef6a4dbe93461c6352741acfc9439a22668f46dd725fcf80fb37d93c88e9d11\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d3f23f4ffaeac5c8606def100262484a45cd86296beb91735215c57955e45067\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://df874a11e199cea6058abe0e5bb63faf7246199fa02093f05720bc41dbb7a81a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://df874a11e199cea6058abe0e5bb63faf7246199fa02093f05720bc41dbb7a81a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:52Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.873043 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"270ac8da-b130-4f4f-bbf7-655229b64446\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:52:31Z\\\",\\\"message\\\":\\\"art network controller: failed to start default network controller: unable to create admin network policy controller, err: could not add Event Handler for anpInformer during admin network policy controller initialization, handler {0x1fcc6e0 0x1fcc3c0 0x1fcc360} was not added to shared informer because it has stopped already, failed to start node network controller: failed to start default node network controller: failed to set node crc annotations: Internal error occurred: failed calling webhook \\\\\\\"node.network-node-identity.openshift.io\\\\\\\": failed to call webhook: Post \\\\\\\"https://127.0.0.1:9743/node?timeout=10s\\\\\\\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:31Z is after 2025-08-24T17:21:41Z]\\\\nI1011 03:52:31.896403 6969 obj_retry.go:303] Retry object setup: *v1.Pod openshift-multus/multus-6xvgx\\\\nI1011 03:52:31.896411 6969 obj_retry.go:365] Adding new object: *v1.Pod openshift-multus/multus-6xvgx\\\\nI1011 03:52:31.896420 6969 ovn.go:134] Ensuring zone local for Pod openshift-multus/multus-6xvgx in node crc\\\\nI1011 03:52:31.896427 6969 obj_retry.go:386] Retry successful for *v1.Pod openshift-multus/multus-6xvgx after 0 failed attempt(s)\\\\nI1011 03:52:31.896434 6969 default_network_controller.go:776] Recording success event on pod openshift-multus/multus-6xvgx\\\\nI1011 03:52:31.8\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:52:30Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lw6dx_openshift-ovn-kubernetes(270ac8da-b130-4f4f-bbf7-655229b64446)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-v8bsk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-lw6dx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:52Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.885311 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.885383 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.885449 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.885478 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.885498 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:52Z","lastTransitionTime":"2025-10-11T03:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.893376 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-p24x8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d22c953c-0315-406e-9c5d-f1759837750e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-6glk9\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:47Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-p24x8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:52Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.920020 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1082af45df3167dfcd11c59d665c17f6fb43fc4c2282e114b67fb8a43f272f6c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a6e8620e12a3aff912c2cfaec7a60b5352e938b47c35b619e162168f34bac526\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:52Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.943254 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:52Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.961547 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-jw555" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"84b4af1b-44af-4ae1-9687-7790849c7d56\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://42f41dee49e71328440037c879d0bba84bffd411b01185ed1b6f3ff3da39e625\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-95dhf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-jw555\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:52Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.983953 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-6xvgx" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"1db525c3-0479-4585-b107-a448f2b94ffd\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:52:22Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://83b1e937441999be327ba6a58c849e9c04956f5cea84da9f8f4aa476e1e79c76\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2025-10-11T03:52:21Z\\\",\\\"message\\\":\\\"2025-10-11T03:51:35+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_74aecb9d-f601-43f8-b48e-632d0155dabc\\\\n2025-10-11T03:51:35+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_74aecb9d-f601-43f8-b48e-632d0155dabc to /host/opt/cni/bin/\\\\n2025-10-11T03:51:36Z [verbose] multus-daemon started\\\\n2025-10-11T03:51:36Z [verbose] Readiness Indicator file check\\\\n2025-10-11T03:52:21Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:52:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-vqqw5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-6xvgx\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:52Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.988130 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.988176 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.988192 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.988216 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:52 crc kubenswrapper[4967]: I1011 03:52:52.988233 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:52Z","lastTransitionTime":"2025-10-11T03:52:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.021867 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-etcd/etcd-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ffe506cf-cbe5-415e-9154-8441ba1d7daa\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7dcd82046c3ffedc54f8b0c1eb37f34a3aae875a3f65161d7f4e853eb870ce69\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8d29662d8a0b3ffb2bdf57eb80188c44de486750c253dd2d5ddbd3fa02afa636\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce9c8d6848994f57e9210a3e9036a7937b450d3c9d4173998659f83928710cd1\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd/\\\",\\\"name\\\":\\\"log-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://df3dd6156b50391db1ddd3cbbbd612c6098d1fba52d262a2228a00c845ecedc8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-rev\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/lib/etcd\\\",\\\"name\\\":\\\"data-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ce68ddd9c5daad7464e3097155d48c8f549cd6a64efd1060777e6ae115a6a82d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcdctl\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/manifests\\\",\\\"name\\\":\\\"static-pod-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/lib/etcd/\\\",\\\"name\\\":\\\"data-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3219cb3ab5a69e602ea57669662086cd09c4afd906fdf5945717666b1ed852c5\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/etcd\\\",\\\"name\\\":\\\"log-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-ensure-env-vars\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://aff5514e2c128f3cf87f6aaedc3e019060ca5cf86a17088f9a9a9b732d0d8e8f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}},{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"etcd-resources-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://3b3f118238e32c485cb55dd378489dedfcf4f846148c4bc1e54e4f17ce2275bb\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/usr/local/bin\\\",\\\"name\\\":\\\"usr-local-bin\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-etcd\"/\"etcd-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.040229 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d0dd7369c8cc48f34caae6b273d6a4d5f27cd11e4a42008144242a0aca948489\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.057172 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.072888 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:32Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://4fdef2e4c6498485b31a5deaf72a2d200612f4457ba8a046921b5fc629b243fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:32Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.090634 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.090712 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.090738 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.090771 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.090793 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:53Z","lastTransitionTime":"2025-10-11T03:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.097159 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f89f20ee-089e-446a-85c4-933aacdb1eba\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:40Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://257b2303215c18aa4500478fbfa39f9e1b95678c6d39ea51e4324c58f51dbd30\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:40Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://eea3c9cb63127599137b9de6bedf417dac7d9e33a5b37c217343041ab0c770ff\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c92258b2b6c6c5028f1e72c0662a19ec11af7b592ea52e50983217716ac5d544\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:35Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:35Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7ae174a489ff0d703529715c18334f2e0bf3a342a3906bcbd8cb1738f9e3a90a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://4d308cfc441d8ecdb7379916503f338eb3b4fa9ab882b5a1ca6b52c019c23036\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:37Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://723d9d5438117e4e48348952d16be0dbb1564702f8b2b2f19f00e6f664a2c573\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:38Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e73fc0d0ae8aef2326aaab33942ae263829c6ab20f959a0468b736bf80003ac3\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:39Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:39Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-t5zwk\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-zwjmt\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.113153 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-48c4c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5187b2c2-8215-4c7c-8906-8f3d275415a0\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:36Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:38Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ad406c24472c9ef7da82485a0600cf54f9002b83c0770c2ac86eca0f8779a870\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:37Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-m7bw8\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:36Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-48c4c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.131696 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"844146aa-36ac-4bee-8015-0d9850d94c4d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:46Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:47Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3cafb41d365e674a55638718df4e540ff77b6bd8e0d2d94bdb480b6617c0942f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ef8f8a8bafbf88541375edb7c1b2134dec31a0d5cd46462a3271d24131bed4c9\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:46Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-r5jll\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:46Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-hs9vz\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.151894 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"47e69eed-ac05-4d04-a290-580433d14207\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7fc96d2d9f48b01b4c4f1f4bc3e058427ffc11b3b73c0ad8a89dbd690a098b67\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://e84c5697ab006ad26c4ebc739bf09997d679c605f84ffe355d30f2b32f4f0e54\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://e84c5697ab006ad26c4ebc739bf09997d679c605f84ffe355d30f2b32f4f0e54\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:13Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.174550 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"e5b43ac7-8e05-48ae-b179-369511b173ff\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:51Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://3d638f32c5784921dbb905323c2e0768379533cce572400402771ebe11ff504f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://26b905095ebfca128bdfd636173bf60583edf9a325e066d661adbdfec980bf9f\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d1444900b2099b98ca09513b4058f2dc4e800416685f5477b3283e301a4f815e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://21500ea1520728049a44298e8acce4b6daa98bcf613d3d0b756858556b53764e\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1450d73d1097928e9a94037d7c1d91a6c48de569724fcf2c48681498adfced42\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"\\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI1011 03:51:31.906207 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nW1011 03:51:31.906229 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906233 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW1011 03:51:31.906237 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nI1011 03:51:31.906233 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW1011 03:51:31.906241 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW1011 03:51:31.906253 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW1011 03:51:31.906256 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nI1011 03:51:31.910364 1 requestheader_controller.go:172] Starting RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910390 1 shared_informer.go:313] Waiting for caches to sync for RequestHeaderAuthRequestController\\\\nI1011 03:51:31.910395 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\\\\"\\\\nI1011 03:51:31.910411 1 shared_informer.go:313] Waiting for caches to sync for client-ca::kube-system::extension-apiserver-authentication::client-ca-file\\\\nI1011 03:51:31.910485 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"WatchListClient\\\\\\\" enabled=false\\\\nI1011 03:51:31.918427 1 envvar.go:172] \\\\\\\"Feature gate default state\\\\\\\" feature=\\\\\\\"InformerResourceVersion\\\\\\\" enabled=false\\\\nI1011 03:51:31.918381 1 configmap_cafile_content.go:205] \\\\\\\"Starting controller\\\\\\\" name=\\\\\\\"client-ca::kube-system::extension-apiserver-authentication::requestheader-client-ca-file\\\\\\\"\\\\nF1011 03:51:31.918495 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:33Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://abed60b2a20b982afb9e9eab57ae34c8bd54c6e723f686b85b658b34b6102181\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:15Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://40b523933534b8142ecdb49735b46ad35446e82fa1e14ccc6cb3a22ddf40dc39\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.190584 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"8f8e3087-41b5-4cb6-9857-73a3702c055f\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:12Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b01603181c7d6005705998a457b0040c682683a90ef782f016ce5c9aaae8c00d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://be0d3cc3a3a52a54783ef5d59b639bef66b28b6d8db74b1cfdae666f8e97199c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:13Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://03b93d108e7fec01df02675a3b8cc0af53c06b89f3bd08841cac46c5fc5a7344\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ea055df636c84578edce0dd638484ac54ea0e50d3db79c781e675310ba6af1fc\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:12Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.194993 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.195065 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.195116 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.195146 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.195167 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:53Z","lastTransitionTime":"2025-10-11T03:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.210821 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:31Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.223766 4967 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"97c07678-14be-410c-b61f-498cb49bc960\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:33Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2025-10-11T03:51:34Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://642e21cd09d7f6cd276fc2755e5331774da7ec4fa675e5cf0ef4edf7744d8555\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2025-10-11T03:51:34Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-fr4zj\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2025-10-11T03:51:33Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-sgrv2\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2025-10-11T03:52:53Z is after 2025-08-24T17:21:41Z" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.298249 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.298317 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.298339 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.298364 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.298385 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:53Z","lastTransitionTime":"2025-10-11T03:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.401726 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.401777 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.401790 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.401808 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.401819 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:53Z","lastTransitionTime":"2025-10-11T03:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.504423 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.504492 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.504515 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.504542 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.504562 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:53Z","lastTransitionTime":"2025-10-11T03:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.607391 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.607435 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.607445 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.607462 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.607475 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:53Z","lastTransitionTime":"2025-10-11T03:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.711321 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.711491 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.711521 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.711601 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.711627 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:53Z","lastTransitionTime":"2025-10-11T03:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.814618 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.814745 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.814637 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:53 crc kubenswrapper[4967]: E1011 03:52:53.814833 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:52:53 crc kubenswrapper[4967]: E1011 03:52:53.814937 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:52:53 crc kubenswrapper[4967]: E1011 03:52:53.815137 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.815273 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.815316 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.815337 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.815365 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.815388 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:53Z","lastTransitionTime":"2025-10-11T03:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.815560 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:53 crc kubenswrapper[4967]: E1011 03:52:53.816057 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.918908 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.918959 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.918972 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.918991 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:53 crc kubenswrapper[4967]: I1011 03:52:53.919010 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:53Z","lastTransitionTime":"2025-10-11T03:52:53Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.022645 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.022731 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.022755 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.022785 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.022807 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:54Z","lastTransitionTime":"2025-10-11T03:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.126191 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.126261 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.126281 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.126310 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.126332 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:54Z","lastTransitionTime":"2025-10-11T03:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.228915 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.228980 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.228997 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.229019 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.229037 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:54Z","lastTransitionTime":"2025-10-11T03:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.332426 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.332486 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.332503 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.332525 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.332544 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:54Z","lastTransitionTime":"2025-10-11T03:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.435774 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.435833 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.435852 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.435875 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.435892 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:54Z","lastTransitionTime":"2025-10-11T03:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.538994 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.539061 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.539130 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.539167 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.539189 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:54Z","lastTransitionTime":"2025-10-11T03:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.642930 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.643000 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.643023 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.643053 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.643119 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:54Z","lastTransitionTime":"2025-10-11T03:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.746564 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.746632 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.746652 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.746680 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.746703 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:54Z","lastTransitionTime":"2025-10-11T03:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.850099 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.850163 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.850182 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.850208 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.850226 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:54Z","lastTransitionTime":"2025-10-11T03:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.954658 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.954734 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.954753 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.954780 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:54 crc kubenswrapper[4967]: I1011 03:52:54.954800 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:54Z","lastTransitionTime":"2025-10-11T03:52:54Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.057433 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.057498 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.057514 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.057537 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.057555 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:55Z","lastTransitionTime":"2025-10-11T03:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.160902 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.160962 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.160979 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.161004 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.161025 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:55Z","lastTransitionTime":"2025-10-11T03:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.263874 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.263969 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.264041 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.264118 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.264141 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:55Z","lastTransitionTime":"2025-10-11T03:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.367528 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.367588 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.367605 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.367628 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.367645 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:55Z","lastTransitionTime":"2025-10-11T03:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.470838 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.470917 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.470941 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.470978 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.471002 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:55Z","lastTransitionTime":"2025-10-11T03:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.574322 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.574450 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.574509 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.574540 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.574558 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:55Z","lastTransitionTime":"2025-10-11T03:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.677263 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.677343 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.677366 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.677397 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.677417 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:55Z","lastTransitionTime":"2025-10-11T03:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.780723 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.780790 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.780814 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.780845 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.780873 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:55Z","lastTransitionTime":"2025-10-11T03:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.814271 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.814325 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.814377 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.814271 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:52:55 crc kubenswrapper[4967]: E1011 03:52:55.814493 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:52:55 crc kubenswrapper[4967]: E1011 03:52:55.814568 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:52:55 crc kubenswrapper[4967]: E1011 03:52:55.814707 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:52:55 crc kubenswrapper[4967]: E1011 03:52:55.815065 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.885061 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.885177 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.885203 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.885235 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.885259 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:55Z","lastTransitionTime":"2025-10-11T03:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.988500 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.988770 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.988874 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.988959 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:55 crc kubenswrapper[4967]: I1011 03:52:55.989154 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:55Z","lastTransitionTime":"2025-10-11T03:52:55Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.092905 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.092972 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.092989 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.093013 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.093031 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:56Z","lastTransitionTime":"2025-10-11T03:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.196227 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.196302 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.196324 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.196356 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.196377 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:56Z","lastTransitionTime":"2025-10-11T03:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.299569 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.299635 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.299661 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.299691 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.299708 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:56Z","lastTransitionTime":"2025-10-11T03:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.403168 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.403242 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.403295 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.403323 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.403345 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:56Z","lastTransitionTime":"2025-10-11T03:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.506838 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.506928 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.506953 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.506981 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.507005 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:56Z","lastTransitionTime":"2025-10-11T03:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.610801 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.610912 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.610932 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.611004 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.611023 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:56Z","lastTransitionTime":"2025-10-11T03:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.713583 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.713654 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.713671 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.713697 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.713715 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:56Z","lastTransitionTime":"2025-10-11T03:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.816017 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.816226 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.816256 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.816283 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.816306 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:56Z","lastTransitionTime":"2025-10-11T03:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.919442 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.919511 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.919538 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.919649 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:56 crc kubenswrapper[4967]: I1011 03:52:56.919673 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:56Z","lastTransitionTime":"2025-10-11T03:52:56Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.021718 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.021820 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.021845 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.021872 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.021893 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:57Z","lastTransitionTime":"2025-10-11T03:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.124399 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.128593 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.128623 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.128675 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.128699 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:57Z","lastTransitionTime":"2025-10-11T03:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.232541 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.232606 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.232624 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.232651 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.232670 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:57Z","lastTransitionTime":"2025-10-11T03:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.335719 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.335779 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.335796 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.335820 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.335838 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:57Z","lastTransitionTime":"2025-10-11T03:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.438436 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.438527 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.438544 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.438566 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.438583 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:57Z","lastTransitionTime":"2025-10-11T03:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.542527 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.542588 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.542606 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.542629 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.542647 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:57Z","lastTransitionTime":"2025-10-11T03:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.647175 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.647255 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.647279 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.647310 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.647330 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:57Z","lastTransitionTime":"2025-10-11T03:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.749830 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.749883 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.749899 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.749922 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.749951 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:57Z","lastTransitionTime":"2025-10-11T03:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.813823 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.813862 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.813939 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:57 crc kubenswrapper[4967]: E1011 03:52:57.814298 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.814468 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:52:57 crc kubenswrapper[4967]: E1011 03:52:57.814822 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:52:57 crc kubenswrapper[4967]: E1011 03:52:57.814954 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:52:57 crc kubenswrapper[4967]: E1011 03:52:57.815065 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.815534 4967 scope.go:117] "RemoveContainer" containerID="2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d" Oct 11 03:52:57 crc kubenswrapper[4967]: E1011 03:52:57.815802 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lw6dx_openshift-ovn-kubernetes(270ac8da-b130-4f4f-bbf7-655229b64446)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.853011 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.853125 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.853149 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.853176 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.853198 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:57Z","lastTransitionTime":"2025-10-11T03:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.956367 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.956479 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.956497 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.956523 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:57 crc kubenswrapper[4967]: I1011 03:52:57.956540 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:57Z","lastTransitionTime":"2025-10-11T03:52:57Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.059839 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.059896 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.059932 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.059965 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.059992 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:58Z","lastTransitionTime":"2025-10-11T03:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.163461 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.163532 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.163557 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.163586 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.163607 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:58Z","lastTransitionTime":"2025-10-11T03:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.266501 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.266562 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.266584 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.266615 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.266672 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:58Z","lastTransitionTime":"2025-10-11T03:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.369650 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.369729 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.369752 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.369782 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.369805 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:58Z","lastTransitionTime":"2025-10-11T03:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.472383 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.472498 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.472524 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.472552 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.472573 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:58Z","lastTransitionTime":"2025-10-11T03:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.576416 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.576472 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.576495 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.576522 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.576542 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:58Z","lastTransitionTime":"2025-10-11T03:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.679902 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.679977 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.679999 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.680030 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.680055 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:58Z","lastTransitionTime":"2025-10-11T03:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.783326 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.783394 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.783410 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.783433 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.783448 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:58Z","lastTransitionTime":"2025-10-11T03:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.887365 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.887428 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.887441 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.887462 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.887475 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:58Z","lastTransitionTime":"2025-10-11T03:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.990338 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.990391 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.990406 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.990426 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:58 crc kubenswrapper[4967]: I1011 03:52:58.990437 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:58Z","lastTransitionTime":"2025-10-11T03:52:58Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.094148 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.094220 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.094237 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.094266 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.094284 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:59Z","lastTransitionTime":"2025-10-11T03:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.198146 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.198210 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.198232 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.198257 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.198278 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:59Z","lastTransitionTime":"2025-10-11T03:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.301159 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.301236 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.301258 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.301288 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.301310 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:59Z","lastTransitionTime":"2025-10-11T03:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.404557 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.404645 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.404667 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.404699 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.404720 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:59Z","lastTransitionTime":"2025-10-11T03:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.507959 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.508003 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.508015 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.508031 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.508044 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:59Z","lastTransitionTime":"2025-10-11T03:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.611881 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.611969 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.611992 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.612025 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.612047 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:59Z","lastTransitionTime":"2025-10-11T03:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.715698 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.715749 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.715761 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.715779 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.715793 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:59Z","lastTransitionTime":"2025-10-11T03:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.813939 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.813981 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.814121 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:52:59 crc kubenswrapper[4967]: E1011 03:52:59.814121 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.814206 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:52:59 crc kubenswrapper[4967]: E1011 03:52:59.814273 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:52:59 crc kubenswrapper[4967]: E1011 03:52:59.814451 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:52:59 crc kubenswrapper[4967]: E1011 03:52:59.814608 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.819015 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.819123 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.819161 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.819187 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.819204 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:59Z","lastTransitionTime":"2025-10-11T03:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.928633 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.928697 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.928731 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.928764 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:52:59 crc kubenswrapper[4967]: I1011 03:52:59.928782 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:52:59Z","lastTransitionTime":"2025-10-11T03:52:59Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.033483 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.033559 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.033583 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.033612 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.033634 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:53:00Z","lastTransitionTime":"2025-10-11T03:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.137334 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.137421 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.137447 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.137478 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.137505 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:53:00Z","lastTransitionTime":"2025-10-11T03:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.240985 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.241475 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.241627 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.241779 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.241919 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:53:00Z","lastTransitionTime":"2025-10-11T03:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.345121 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.345549 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.345728 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.345910 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.346140 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:53:00Z","lastTransitionTime":"2025-10-11T03:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.448870 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.449943 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.450223 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.450438 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.450621 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:53:00Z","lastTransitionTime":"2025-10-11T03:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.553952 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.554025 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.554041 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.554062 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.554141 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:53:00Z","lastTransitionTime":"2025-10-11T03:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.656461 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.656503 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.656511 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.656528 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.656538 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:53:00Z","lastTransitionTime":"2025-10-11T03:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.759254 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.759327 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.759350 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.759377 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.759398 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:53:00Z","lastTransitionTime":"2025-10-11T03:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.795037 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.795147 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.795166 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.795192 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.795210 4967 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2025-10-11T03:53:00Z","lastTransitionTime":"2025-10-11T03:53:00Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.874337 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-m8rrp"] Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.875573 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m8rrp" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.878578 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.879009 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.879140 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.879370 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.907541 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=88.907481547 podStartE2EDuration="1m28.907481547s" podCreationTimestamp="2025-10-11 03:51:32 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:00.906305697 +0000 UTC m=+108.869514670" watchObservedRunningTime="2025-10-11 03:53:00.907481547 +0000 UTC m=+108.870690510" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.958661 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=87.958631051 podStartE2EDuration="1m27.958631051s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:00.936246576 +0000 UTC m=+108.899455549" watchObservedRunningTime="2025-10-11 03:53:00.958631051 +0000 UTC m=+108.921840024" Oct 11 03:53:00 crc kubenswrapper[4967]: I1011 03:53:00.994499 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podStartSLOduration=87.994476872 podStartE2EDuration="1m27.994476872s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:00.976256894 +0000 UTC m=+108.939465967" watchObservedRunningTime="2025-10-11 03:53:00.994476872 +0000 UTC m=+108.957685825" Oct 11 03:53:01 crc kubenswrapper[4967]: I1011 03:53:01.029716 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/b8ca32e8-b8bf-4f35-b558-d11237d172ea-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-m8rrp\" (UID: \"b8ca32e8-b8bf-4f35-b558-d11237d172ea\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m8rrp" Oct 11 03:53:01 crc kubenswrapper[4967]: I1011 03:53:01.030038 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/b8ca32e8-b8bf-4f35-b558-d11237d172ea-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-m8rrp\" (UID: \"b8ca32e8-b8bf-4f35-b558-d11237d172ea\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m8rrp" Oct 11 03:53:01 crc kubenswrapper[4967]: I1011 03:53:01.030163 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b8ca32e8-b8bf-4f35-b558-d11237d172ea-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-m8rrp\" (UID: \"b8ca32e8-b8bf-4f35-b558-d11237d172ea\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m8rrp" Oct 11 03:53:01 crc kubenswrapper[4967]: I1011 03:53:01.030235 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b8ca32e8-b8bf-4f35-b558-d11237d172ea-service-ca\") pod \"cluster-version-operator-5c965bbfc6-m8rrp\" (UID: \"b8ca32e8-b8bf-4f35-b558-d11237d172ea\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m8rrp" Oct 11 03:53:01 crc kubenswrapper[4967]: I1011 03:53:01.030309 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8ca32e8-b8bf-4f35-b558-d11237d172ea-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-m8rrp\" (UID: \"b8ca32e8-b8bf-4f35-b558-d11237d172ea\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m8rrp" Oct 11 03:53:01 crc kubenswrapper[4967]: I1011 03:53:01.033030 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=59.033002702 podStartE2EDuration="59.033002702s" podCreationTimestamp="2025-10-11 03:52:02 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:00.99598062 +0000 UTC m=+108.959189593" watchObservedRunningTime="2025-10-11 03:53:01.033002702 +0000 UTC m=+108.996211675" Oct 11 03:53:01 crc kubenswrapper[4967]: I1011 03:53:01.075100 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-jw555" podStartSLOduration=88.075051102 podStartE2EDuration="1m28.075051102s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:01.074879548 +0000 UTC m=+109.038088481" watchObservedRunningTime="2025-10-11 03:53:01.075051102 +0000 UTC m=+109.038260045" Oct 11 03:53:01 crc kubenswrapper[4967]: I1011 03:53:01.096906 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-6xvgx" podStartSLOduration=88.096888603 podStartE2EDuration="1m28.096888603s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:01.096832572 +0000 UTC m=+109.060041495" watchObservedRunningTime="2025-10-11 03:53:01.096888603 +0000 UTC m=+109.060097546" Oct 11 03:53:01 crc kubenswrapper[4967]: I1011 03:53:01.131490 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b8ca32e8-b8bf-4f35-b558-d11237d172ea-service-ca\") pod \"cluster-version-operator-5c965bbfc6-m8rrp\" (UID: \"b8ca32e8-b8bf-4f35-b558-d11237d172ea\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m8rrp" Oct 11 03:53:01 crc kubenswrapper[4967]: I1011 03:53:01.131566 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8ca32e8-b8bf-4f35-b558-d11237d172ea-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-m8rrp\" (UID: \"b8ca32e8-b8bf-4f35-b558-d11237d172ea\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m8rrp" Oct 11 03:53:01 crc kubenswrapper[4967]: I1011 03:53:01.131636 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/b8ca32e8-b8bf-4f35-b558-d11237d172ea-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-m8rrp\" (UID: \"b8ca32e8-b8bf-4f35-b558-d11237d172ea\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m8rrp" Oct 11 03:53:01 crc kubenswrapper[4967]: I1011 03:53:01.131708 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/b8ca32e8-b8bf-4f35-b558-d11237d172ea-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-m8rrp\" (UID: \"b8ca32e8-b8bf-4f35-b558-d11237d172ea\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m8rrp" Oct 11 03:53:01 crc kubenswrapper[4967]: I1011 03:53:01.131765 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b8ca32e8-b8bf-4f35-b558-d11237d172ea-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-m8rrp\" (UID: \"b8ca32e8-b8bf-4f35-b558-d11237d172ea\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m8rrp" Oct 11 03:53:01 crc kubenswrapper[4967]: I1011 03:53:01.132201 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/b8ca32e8-b8bf-4f35-b558-d11237d172ea-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-m8rrp\" (UID: \"b8ca32e8-b8bf-4f35-b558-d11237d172ea\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m8rrp" Oct 11 03:53:01 crc kubenswrapper[4967]: I1011 03:53:01.132217 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/b8ca32e8-b8bf-4f35-b558-d11237d172ea-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-m8rrp\" (UID: \"b8ca32e8-b8bf-4f35-b558-d11237d172ea\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m8rrp" Oct 11 03:53:01 crc kubenswrapper[4967]: I1011 03:53:01.133275 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/b8ca32e8-b8bf-4f35-b558-d11237d172ea-service-ca\") pod \"cluster-version-operator-5c965bbfc6-m8rrp\" (UID: \"b8ca32e8-b8bf-4f35-b558-d11237d172ea\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m8rrp" Oct 11 03:53:01 crc kubenswrapper[4967]: I1011 03:53:01.133966 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=88.133953175 podStartE2EDuration="1m28.133953175s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:01.132840966 +0000 UTC m=+109.096049909" watchObservedRunningTime="2025-10-11 03:53:01.133953175 +0000 UTC m=+109.097162108" Oct 11 03:53:01 crc kubenswrapper[4967]: I1011 03:53:01.139059 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b8ca32e8-b8bf-4f35-b558-d11237d172ea-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-m8rrp\" (UID: \"b8ca32e8-b8bf-4f35-b558-d11237d172ea\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m8rrp" Oct 11 03:53:01 crc kubenswrapper[4967]: I1011 03:53:01.151990 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b8ca32e8-b8bf-4f35-b558-d11237d172ea-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-m8rrp\" (UID: \"b8ca32e8-b8bf-4f35-b558-d11237d172ea\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m8rrp" Oct 11 03:53:01 crc kubenswrapper[4967]: I1011 03:53:01.201804 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m8rrp" Oct 11 03:53:01 crc kubenswrapper[4967]: I1011 03:53:01.261210 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-48c4c" podStartSLOduration=88.261187143 podStartE2EDuration="1m28.261187143s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:01.261154342 +0000 UTC m=+109.224363275" watchObservedRunningTime="2025-10-11 03:53:01.261187143 +0000 UTC m=+109.224396076" Oct 11 03:53:01 crc kubenswrapper[4967]: I1011 03:53:01.261723 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-zwjmt" podStartSLOduration=88.261717667 podStartE2EDuration="1m28.261717667s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:01.248724193 +0000 UTC m=+109.211933126" watchObservedRunningTime="2025-10-11 03:53:01.261717667 +0000 UTC m=+109.224926600" Oct 11 03:53:01 crc kubenswrapper[4967]: I1011 03:53:01.284746 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-hs9vz" podStartSLOduration=88.284726058 podStartE2EDuration="1m28.284726058s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:01.272625277 +0000 UTC m=+109.235834210" watchObservedRunningTime="2025-10-11 03:53:01.284726058 +0000 UTC m=+109.247934991" Oct 11 03:53:01 crc kubenswrapper[4967]: I1011 03:53:01.296529 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=27.29650406 podStartE2EDuration="27.29650406s" podCreationTimestamp="2025-10-11 03:52:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:01.295392632 +0000 UTC m=+109.258601575" watchObservedRunningTime="2025-10-11 03:53:01.29650406 +0000 UTC m=+109.259713023" Oct 11 03:53:01 crc kubenswrapper[4967]: I1011 03:53:01.399281 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m8rrp" event={"ID":"b8ca32e8-b8bf-4f35-b558-d11237d172ea","Type":"ContainerStarted","Data":"1d82aa7575ef521b1fc8dae5c4ad19503febd59fe75c85097699df93b46fdbf4"} Oct 11 03:53:01 crc kubenswrapper[4967]: I1011 03:53:01.399330 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m8rrp" event={"ID":"b8ca32e8-b8bf-4f35-b558-d11237d172ea","Type":"ContainerStarted","Data":"fdd573b1c7deb9350d1469506c6124914f4db0848a4fe2d50f2f26f089720d70"} Oct 11 03:53:01 crc kubenswrapper[4967]: I1011 03:53:01.414325 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-m8rrp" podStartSLOduration=88.414303886 podStartE2EDuration="1m28.414303886s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:01.413008283 +0000 UTC m=+109.376217256" watchObservedRunningTime="2025-10-11 03:53:01.414303886 +0000 UTC m=+109.377512829" Oct 11 03:53:01 crc kubenswrapper[4967]: I1011 03:53:01.814747 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:53:01 crc kubenswrapper[4967]: I1011 03:53:01.814786 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:53:01 crc kubenswrapper[4967]: I1011 03:53:01.814879 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:53:01 crc kubenswrapper[4967]: E1011 03:53:01.815050 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:53:01 crc kubenswrapper[4967]: E1011 03:53:01.815227 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:53:01 crc kubenswrapper[4967]: E1011 03:53:01.815364 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:53:01 crc kubenswrapper[4967]: I1011 03:53:01.814780 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:53:01 crc kubenswrapper[4967]: E1011 03:53:01.816465 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:53:03 crc kubenswrapper[4967]: I1011 03:53:03.814788 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:53:03 crc kubenswrapper[4967]: I1011 03:53:03.814868 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:53:03 crc kubenswrapper[4967]: I1011 03:53:03.814829 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:53:03 crc kubenswrapper[4967]: E1011 03:53:03.814955 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:53:03 crc kubenswrapper[4967]: E1011 03:53:03.815777 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:53:03 crc kubenswrapper[4967]: I1011 03:53:03.815690 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:53:03 crc kubenswrapper[4967]: E1011 03:53:03.815843 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:53:03 crc kubenswrapper[4967]: E1011 03:53:03.815934 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:53:05 crc kubenswrapper[4967]: I1011 03:53:05.814672 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:53:05 crc kubenswrapper[4967]: E1011 03:53:05.814867 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:53:05 crc kubenswrapper[4967]: I1011 03:53:05.814916 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:53:05 crc kubenswrapper[4967]: I1011 03:53:05.815111 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:53:05 crc kubenswrapper[4967]: E1011 03:53:05.815328 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:53:05 crc kubenswrapper[4967]: I1011 03:53:05.815468 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:53:05 crc kubenswrapper[4967]: E1011 03:53:05.815602 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:53:05 crc kubenswrapper[4967]: E1011 03:53:05.815731 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:53:07 crc kubenswrapper[4967]: I1011 03:53:07.419787 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6xvgx_1db525c3-0479-4585-b107-a448f2b94ffd/kube-multus/1.log" Oct 11 03:53:07 crc kubenswrapper[4967]: I1011 03:53:07.420450 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6xvgx_1db525c3-0479-4585-b107-a448f2b94ffd/kube-multus/0.log" Oct 11 03:53:07 crc kubenswrapper[4967]: I1011 03:53:07.420514 4967 generic.go:334] "Generic (PLEG): container finished" podID="1db525c3-0479-4585-b107-a448f2b94ffd" containerID="83b1e937441999be327ba6a58c849e9c04956f5cea84da9f8f4aa476e1e79c76" exitCode=1 Oct 11 03:53:07 crc kubenswrapper[4967]: I1011 03:53:07.420561 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6xvgx" event={"ID":"1db525c3-0479-4585-b107-a448f2b94ffd","Type":"ContainerDied","Data":"83b1e937441999be327ba6a58c849e9c04956f5cea84da9f8f4aa476e1e79c76"} Oct 11 03:53:07 crc kubenswrapper[4967]: I1011 03:53:07.420630 4967 scope.go:117] "RemoveContainer" containerID="a82f45ad38dd837669e2c391fcefba7840f037a6165572ac079beb8ee7776fb0" Oct 11 03:53:07 crc kubenswrapper[4967]: I1011 03:53:07.421145 4967 scope.go:117] "RemoveContainer" containerID="83b1e937441999be327ba6a58c849e9c04956f5cea84da9f8f4aa476e1e79c76" Oct 11 03:53:07 crc kubenswrapper[4967]: E1011 03:53:07.421392 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-6xvgx_openshift-multus(1db525c3-0479-4585-b107-a448f2b94ffd)\"" pod="openshift-multus/multus-6xvgx" podUID="1db525c3-0479-4585-b107-a448f2b94ffd" Oct 11 03:53:07 crc kubenswrapper[4967]: I1011 03:53:07.814276 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:53:07 crc kubenswrapper[4967]: I1011 03:53:07.814377 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:53:07 crc kubenswrapper[4967]: I1011 03:53:07.814466 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:53:07 crc kubenswrapper[4967]: E1011 03:53:07.814472 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:53:07 crc kubenswrapper[4967]: I1011 03:53:07.814524 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:53:07 crc kubenswrapper[4967]: E1011 03:53:07.814707 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:53:07 crc kubenswrapper[4967]: E1011 03:53:07.814847 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:53:07 crc kubenswrapper[4967]: E1011 03:53:07.814924 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:53:08 crc kubenswrapper[4967]: I1011 03:53:08.425311 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6xvgx_1db525c3-0479-4585-b107-a448f2b94ffd/kube-multus/1.log" Oct 11 03:53:09 crc kubenswrapper[4967]: I1011 03:53:09.814159 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:53:09 crc kubenswrapper[4967]: I1011 03:53:09.814208 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:53:09 crc kubenswrapper[4967]: I1011 03:53:09.814184 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:53:09 crc kubenswrapper[4967]: I1011 03:53:09.814161 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:53:09 crc kubenswrapper[4967]: E1011 03:53:09.814310 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:53:09 crc kubenswrapper[4967]: E1011 03:53:09.814450 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:53:09 crc kubenswrapper[4967]: E1011 03:53:09.814639 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:53:09 crc kubenswrapper[4967]: E1011 03:53:09.814684 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:53:10 crc kubenswrapper[4967]: I1011 03:53:10.816146 4967 scope.go:117] "RemoveContainer" containerID="2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d" Oct 11 03:53:10 crc kubenswrapper[4967]: E1011 03:53:10.816495 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-lw6dx_openshift-ovn-kubernetes(270ac8da-b130-4f4f-bbf7-655229b64446)\"" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" Oct 11 03:53:11 crc kubenswrapper[4967]: I1011 03:53:11.814195 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:53:11 crc kubenswrapper[4967]: I1011 03:53:11.814248 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:53:11 crc kubenswrapper[4967]: I1011 03:53:11.814200 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:53:11 crc kubenswrapper[4967]: I1011 03:53:11.814300 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:53:11 crc kubenswrapper[4967]: E1011 03:53:11.814353 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:53:11 crc kubenswrapper[4967]: E1011 03:53:11.814438 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:53:11 crc kubenswrapper[4967]: E1011 03:53:11.814570 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:53:11 crc kubenswrapper[4967]: E1011 03:53:11.814676 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:53:12 crc kubenswrapper[4967]: E1011 03:53:12.794402 4967 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Oct 11 03:53:12 crc kubenswrapper[4967]: E1011 03:53:12.908575 4967 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 11 03:53:13 crc kubenswrapper[4967]: I1011 03:53:13.814119 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:53:13 crc kubenswrapper[4967]: I1011 03:53:13.814205 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:53:13 crc kubenswrapper[4967]: I1011 03:53:13.814212 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:53:13 crc kubenswrapper[4967]: I1011 03:53:13.814134 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:53:13 crc kubenswrapper[4967]: E1011 03:53:13.814334 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:53:13 crc kubenswrapper[4967]: E1011 03:53:13.814510 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:53:13 crc kubenswrapper[4967]: E1011 03:53:13.814653 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:53:13 crc kubenswrapper[4967]: E1011 03:53:13.814773 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:53:15 crc kubenswrapper[4967]: I1011 03:53:15.814469 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:53:15 crc kubenswrapper[4967]: I1011 03:53:15.814549 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:53:15 crc kubenswrapper[4967]: I1011 03:53:15.814473 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:53:15 crc kubenswrapper[4967]: I1011 03:53:15.814504 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:53:15 crc kubenswrapper[4967]: E1011 03:53:15.814679 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:53:15 crc kubenswrapper[4967]: E1011 03:53:15.814792 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:53:15 crc kubenswrapper[4967]: E1011 03:53:15.814916 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:53:15 crc kubenswrapper[4967]: E1011 03:53:15.815122 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:53:17 crc kubenswrapper[4967]: I1011 03:53:17.814378 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:53:17 crc kubenswrapper[4967]: I1011 03:53:17.814427 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:53:17 crc kubenswrapper[4967]: I1011 03:53:17.814485 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:53:17 crc kubenswrapper[4967]: E1011 03:53:17.814606 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:53:17 crc kubenswrapper[4967]: I1011 03:53:17.814662 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:53:17 crc kubenswrapper[4967]: E1011 03:53:17.814903 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:53:17 crc kubenswrapper[4967]: E1011 03:53:17.815042 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:53:17 crc kubenswrapper[4967]: E1011 03:53:17.815234 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:53:17 crc kubenswrapper[4967]: E1011 03:53:17.910220 4967 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 11 03:53:18 crc kubenswrapper[4967]: I1011 03:53:18.814852 4967 scope.go:117] "RemoveContainer" containerID="83b1e937441999be327ba6a58c849e9c04956f5cea84da9f8f4aa476e1e79c76" Oct 11 03:53:19 crc kubenswrapper[4967]: I1011 03:53:19.466658 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6xvgx_1db525c3-0479-4585-b107-a448f2b94ffd/kube-multus/1.log" Oct 11 03:53:19 crc kubenswrapper[4967]: I1011 03:53:19.466720 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6xvgx" event={"ID":"1db525c3-0479-4585-b107-a448f2b94ffd","Type":"ContainerStarted","Data":"6303c6895d8a61d2d88fe615b28d6b87e1180831b83ef4f6e1019a7d186554fc"} Oct 11 03:53:19 crc kubenswrapper[4967]: I1011 03:53:19.813960 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:53:19 crc kubenswrapper[4967]: I1011 03:53:19.813979 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:53:19 crc kubenswrapper[4967]: I1011 03:53:19.814013 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:53:19 crc kubenswrapper[4967]: I1011 03:53:19.814002 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:53:19 crc kubenswrapper[4967]: E1011 03:53:19.815452 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:53:19 crc kubenswrapper[4967]: E1011 03:53:19.815684 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:53:19 crc kubenswrapper[4967]: E1011 03:53:19.815797 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:53:19 crc kubenswrapper[4967]: E1011 03:53:19.815906 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:53:21 crc kubenswrapper[4967]: I1011 03:53:21.814387 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:53:21 crc kubenswrapper[4967]: I1011 03:53:21.814432 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:53:21 crc kubenswrapper[4967]: I1011 03:53:21.814475 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:53:21 crc kubenswrapper[4967]: E1011 03:53:21.814521 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:53:21 crc kubenswrapper[4967]: I1011 03:53:21.814571 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:53:21 crc kubenswrapper[4967]: E1011 03:53:21.814683 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:53:21 crc kubenswrapper[4967]: E1011 03:53:21.814819 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:53:21 crc kubenswrapper[4967]: E1011 03:53:21.814942 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:53:22 crc kubenswrapper[4967]: I1011 03:53:22.817771 4967 scope.go:117] "RemoveContainer" containerID="2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d" Oct 11 03:53:22 crc kubenswrapper[4967]: E1011 03:53:22.911427 4967 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Oct 11 03:53:23 crc kubenswrapper[4967]: I1011 03:53:23.482768 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lw6dx_270ac8da-b130-4f4f-bbf7-655229b64446/ovnkube-controller/3.log" Oct 11 03:53:23 crc kubenswrapper[4967]: I1011 03:53:23.485383 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" event={"ID":"270ac8da-b130-4f4f-bbf7-655229b64446","Type":"ContainerStarted","Data":"46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1"} Oct 11 03:53:23 crc kubenswrapper[4967]: I1011 03:53:23.485721 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:53:23 crc kubenswrapper[4967]: I1011 03:53:23.515697 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" podStartSLOduration=110.515677538 podStartE2EDuration="1m50.515677538s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:23.51536375 +0000 UTC m=+131.478572693" watchObservedRunningTime="2025-10-11 03:53:23.515677538 +0000 UTC m=+131.478886501" Oct 11 03:53:23 crc kubenswrapper[4967]: I1011 03:53:23.773754 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-p24x8"] Oct 11 03:53:23 crc kubenswrapper[4967]: I1011 03:53:23.773891 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:53:23 crc kubenswrapper[4967]: E1011 03:53:23.774025 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:53:23 crc kubenswrapper[4967]: I1011 03:53:23.814601 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:53:23 crc kubenswrapper[4967]: I1011 03:53:23.814681 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:53:23 crc kubenswrapper[4967]: E1011 03:53:23.814762 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:53:23 crc kubenswrapper[4967]: I1011 03:53:23.814832 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:53:23 crc kubenswrapper[4967]: E1011 03:53:23.815050 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:53:23 crc kubenswrapper[4967]: E1011 03:53:23.815279 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:53:25 crc kubenswrapper[4967]: I1011 03:53:25.813780 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:53:25 crc kubenswrapper[4967]: E1011 03:53:25.814239 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:53:25 crc kubenswrapper[4967]: I1011 03:53:25.813887 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:53:25 crc kubenswrapper[4967]: I1011 03:53:25.813789 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:53:25 crc kubenswrapper[4967]: E1011 03:53:25.814334 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:53:25 crc kubenswrapper[4967]: I1011 03:53:25.813932 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:53:25 crc kubenswrapper[4967]: E1011 03:53:25.814703 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:53:25 crc kubenswrapper[4967]: E1011 03:53:25.814507 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:53:27 crc kubenswrapper[4967]: I1011 03:53:27.814961 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:53:27 crc kubenswrapper[4967]: I1011 03:53:27.815000 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:53:27 crc kubenswrapper[4967]: I1011 03:53:27.815116 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:53:27 crc kubenswrapper[4967]: E1011 03:53:27.815221 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Oct 11 03:53:27 crc kubenswrapper[4967]: I1011 03:53:27.815225 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:53:27 crc kubenswrapper[4967]: E1011 03:53:27.815600 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-p24x8" podUID="d22c953c-0315-406e-9c5d-f1759837750e" Oct 11 03:53:27 crc kubenswrapper[4967]: E1011 03:53:27.815843 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Oct 11 03:53:27 crc kubenswrapper[4967]: E1011 03:53:27.816061 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Oct 11 03:53:29 crc kubenswrapper[4967]: I1011 03:53:29.814529 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:53:29 crc kubenswrapper[4967]: I1011 03:53:29.814573 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:53:29 crc kubenswrapper[4967]: I1011 03:53:29.814624 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:53:29 crc kubenswrapper[4967]: I1011 03:53:29.814542 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:53:29 crc kubenswrapper[4967]: I1011 03:53:29.816987 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Oct 11 03:53:29 crc kubenswrapper[4967]: I1011 03:53:29.817335 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Oct 11 03:53:29 crc kubenswrapper[4967]: I1011 03:53:29.817433 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Oct 11 03:53:29 crc kubenswrapper[4967]: I1011 03:53:29.817915 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Oct 11 03:53:29 crc kubenswrapper[4967]: I1011 03:53:29.818393 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Oct 11 03:53:29 crc kubenswrapper[4967]: I1011 03:53:29.818593 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.334303 4967 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.401194 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-92sz4"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.401844 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-92sz4" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.405262 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.405769 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.406137 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.406899 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.410192 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-mrsjk"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.411132 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-mrsjk" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.412147 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-nxdtr"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.412847 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.413376 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.413416 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.419640 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.420718 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.420917 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.422119 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.422399 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.424181 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.424257 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.424346 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.424442 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.424500 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wtn79"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.424510 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.424593 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.425196 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wtn79" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.424618 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.424665 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.427632 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-vwp4c"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.428491 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwp4c" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.428660 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.428849 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.429017 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.432900 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.433166 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ws79r"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.439448 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bjc5w"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.440176 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-dtj88"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.441920 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-dtj88" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.442839 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ws79r" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.443295 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bjc5w" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.448337 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-rt5vd"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.452581 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.452921 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-rt5vd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.467897 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.468221 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.468298 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.468603 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.468765 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.468975 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.469029 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.469061 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.469243 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.469776 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.470488 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.470679 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.470742 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-qn4g8"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.470833 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.471531 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kmpsc"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.472012 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-9jjwn"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.472037 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kmpsc" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.472085 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.472157 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-qn4g8" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.472651 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.472752 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.472790 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.472924 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.472940 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.473111 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-ws6r9"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.473618 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ws6r9" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.473872 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-9jjwn" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.475459 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.475682 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.475807 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.476089 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.477337 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.477642 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.477682 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.477807 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.477903 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7sxxj"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.477960 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.478112 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.478154 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.478289 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.478395 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-7sxxj" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.478447 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.478520 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.478798 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.478835 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.478834 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6sjfd"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.479377 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.484041 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-csl6r"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.484909 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-csl6r" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.485058 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rw8ng"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.485686 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rw8ng" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.486647 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1c77fe9a-d2e1-465b-8600-9f7f2d1fa206-client-ca\") pod \"controller-manager-879f6c89f-92sz4\" (UID: \"1c77fe9a-d2e1-465b-8600-9f7f2d1fa206\") " pod="openshift-controller-manager/controller-manager-879f6c89f-92sz4" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.486685 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c77fe9a-d2e1-465b-8600-9f7f2d1fa206-config\") pod \"controller-manager-879f6c89f-92sz4\" (UID: \"1c77fe9a-d2e1-465b-8600-9f7f2d1fa206\") " pod="openshift-controller-manager/controller-manager-879f6c89f-92sz4" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.486728 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/454804e7-0d4d-4536-8544-87a5b6368f70-etcd-client\") pod \"apiserver-7bbb656c7d-nmzrx\" (UID: \"454804e7-0d4d-4536-8544-87a5b6368f70\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.486751 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9w2rs\" (UniqueName: \"kubernetes.io/projected/9b3efa80-8370-4353-b688-4701c4deec8f-kube-api-access-9w2rs\") pod \"machine-api-operator-5694c8668f-mrsjk\" (UID: \"9b3efa80-8370-4353-b688-4701c4deec8f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mrsjk" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.486784 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/454804e7-0d4d-4536-8544-87a5b6368f70-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-nmzrx\" (UID: \"454804e7-0d4d-4536-8544-87a5b6368f70\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.486807 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/454804e7-0d4d-4536-8544-87a5b6368f70-encryption-config\") pod \"apiserver-7bbb656c7d-nmzrx\" (UID: \"454804e7-0d4d-4536-8544-87a5b6368f70\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.486828 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efe9c94d-8f97-4274-a494-a5ecf5b88cec-config\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.486847 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/efe9c94d-8f97-4274-a494-a5ecf5b88cec-image-import-ca\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.486871 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/454804e7-0d4d-4536-8544-87a5b6368f70-audit-dir\") pod \"apiserver-7bbb656c7d-nmzrx\" (UID: \"454804e7-0d4d-4536-8544-87a5b6368f70\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.486891 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b3efa80-8370-4353-b688-4701c4deec8f-config\") pod \"machine-api-operator-5694c8668f-mrsjk\" (UID: \"9b3efa80-8370-4353-b688-4701c4deec8f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mrsjk" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.486911 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1c77fe9a-d2e1-465b-8600-9f7f2d1fa206-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-92sz4\" (UID: \"1c77fe9a-d2e1-465b-8600-9f7f2d1fa206\") " pod="openshift-controller-manager/controller-manager-879f6c89f-92sz4" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.486932 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hk288\" (UniqueName: \"kubernetes.io/projected/efe9c94d-8f97-4274-a494-a5ecf5b88cec-kube-api-access-hk288\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.486955 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8vnp\" (UniqueName: \"kubernetes.io/projected/1c77fe9a-d2e1-465b-8600-9f7f2d1fa206-kube-api-access-j8vnp\") pod \"controller-manager-879f6c89f-92sz4\" (UID: \"1c77fe9a-d2e1-465b-8600-9f7f2d1fa206\") " pod="openshift-controller-manager/controller-manager-879f6c89f-92sz4" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.486976 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/efe9c94d-8f97-4274-a494-a5ecf5b88cec-serving-cert\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.487003 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/454804e7-0d4d-4536-8544-87a5b6368f70-serving-cert\") pod \"apiserver-7bbb656c7d-nmzrx\" (UID: \"454804e7-0d4d-4536-8544-87a5b6368f70\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.487023 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9b3efa80-8370-4353-b688-4701c4deec8f-images\") pod \"machine-api-operator-5694c8668f-mrsjk\" (UID: \"9b3efa80-8370-4353-b688-4701c4deec8f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mrsjk" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.487046 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/454804e7-0d4d-4536-8544-87a5b6368f70-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-nmzrx\" (UID: \"454804e7-0d4d-4536-8544-87a5b6368f70\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.487067 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qg482\" (UniqueName: \"kubernetes.io/projected/454804e7-0d4d-4536-8544-87a5b6368f70-kube-api-access-qg482\") pod \"apiserver-7bbb656c7d-nmzrx\" (UID: \"454804e7-0d4d-4536-8544-87a5b6368f70\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.487267 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/454804e7-0d4d-4536-8544-87a5b6368f70-audit-policies\") pod \"apiserver-7bbb656c7d-nmzrx\" (UID: \"454804e7-0d4d-4536-8544-87a5b6368f70\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.487287 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/efe9c94d-8f97-4274-a494-a5ecf5b88cec-etcd-serving-ca\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.487312 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/9b3efa80-8370-4353-b688-4701c4deec8f-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-mrsjk\" (UID: \"9b3efa80-8370-4353-b688-4701c4deec8f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mrsjk" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.487334 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c77fe9a-d2e1-465b-8600-9f7f2d1fa206-serving-cert\") pod \"controller-manager-879f6c89f-92sz4\" (UID: \"1c77fe9a-d2e1-465b-8600-9f7f2d1fa206\") " pod="openshift-controller-manager/controller-manager-879f6c89f-92sz4" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.487354 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/efe9c94d-8f97-4274-a494-a5ecf5b88cec-etcd-client\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.487372 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/efe9c94d-8f97-4274-a494-a5ecf5b88cec-trusted-ca-bundle\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.487398 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/efe9c94d-8f97-4274-a494-a5ecf5b88cec-node-pullsecrets\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.487419 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/efe9c94d-8f97-4274-a494-a5ecf5b88cec-encryption-config\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.487459 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/efe9c94d-8f97-4274-a494-a5ecf5b88cec-audit-dir\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.487478 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/efe9c94d-8f97-4274-a494-a5ecf5b88cec-audit\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.487731 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gwrxz"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.488231 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-cwz9p"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.488687 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-n6qrf"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.490397 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gwrxz" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.491140 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cwz9p" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.493286 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.494483 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.494788 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-cp2n5"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.495190 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.495364 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tr7z7"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.495695 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rmkqj"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.496375 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.496780 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.497089 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.498759 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.499036 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.499597 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.500485 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rv6n2"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.501381 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-gbqfq"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.501666 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.507330 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.507464 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.507531 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.507531 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.521413 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.522699 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.522892 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.524380 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n6qrf" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.525154 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cp2n5" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.525954 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bxjls"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.528205 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rmkqj" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.528869 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7nk8v"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.529182 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.535434 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rv6n2" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.536655 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.538651 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.538735 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.540215 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bxjls" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.540319 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.540397 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.540475 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.540551 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.541407 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gbqfq" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.541568 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.543258 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7nk8v" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.543339 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v7xvv"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.544242 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnjpj"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.544871 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v7xvv" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.544977 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.545032 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-2fvpq"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.545052 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tr7z7" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.545130 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnjpj" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.545139 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.545478 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-g4q8m"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.546148 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-2fvpq" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.546168 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hk2cx"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.546524 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.546620 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.546697 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.546785 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.548974 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.548992 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.549156 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.549576 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-g4q8m" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.551250 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.552387 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.553917 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.554480 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.554523 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.558299 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.561721 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mhnj2"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.562576 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mhnj2" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.562591 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-49rxs"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.563131 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-49rxs" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.563441 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.563621 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-9c49l"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.563980 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-9c49l" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.565882 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.566049 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-4xkx2"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.566830 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335905-kdpld"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.567131 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.567364 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-kdpld" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.567421 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.567534 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-4xkx2" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.567583 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.567759 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.568552 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.568830 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-mrsjk"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.570155 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kmpsc"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.575585 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.580526 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-rt5vd"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.580586 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-nxdtr"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.584556 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-dtj88"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.584601 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bxjls"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.584636 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ws79r"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.587607 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.587769 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gwrxz"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.588673 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c77fe9a-d2e1-465b-8600-9f7f2d1fa206-config\") pod \"controller-manager-879f6c89f-92sz4\" (UID: \"1c77fe9a-d2e1-465b-8600-9f7f2d1fa206\") " pod="openshift-controller-manager/controller-manager-879f6c89f-92sz4" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.588713 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kp59m\" (UniqueName: \"kubernetes.io/projected/092da80d-ee9c-4e88-b3e2-58bb3dd77f9b-kube-api-access-kp59m\") pod \"openshift-controller-manager-operator-756b6f6bc6-tr7z7\" (UID: \"092da80d-ee9c-4e88-b3e2-58bb3dd77f9b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tr7z7" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.588739 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9w2rs\" (UniqueName: \"kubernetes.io/projected/9b3efa80-8370-4353-b688-4701c4deec8f-kube-api-access-9w2rs\") pod \"machine-api-operator-5694c8668f-mrsjk\" (UID: \"9b3efa80-8370-4353-b688-4701c4deec8f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mrsjk" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.588782 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/49347426-bca5-4fde-8b87-21dc8fc486d8-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-kmpsc\" (UID: \"49347426-bca5-4fde-8b87-21dc8fc486d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kmpsc" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.588804 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rwhh8\" (UniqueName: \"kubernetes.io/projected/3bc0bb1d-5b22-4e74-8aa8-6554b77a2b73-kube-api-access-rwhh8\") pod \"authentication-operator-69f744f599-qn4g8\" (UID: \"3bc0bb1d-5b22-4e74-8aa8-6554b77a2b73\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qn4g8" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.588836 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/454804e7-0d4d-4536-8544-87a5b6368f70-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-nmzrx\" (UID: \"454804e7-0d4d-4536-8544-87a5b6368f70\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.588881 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/efe9c94d-8f97-4274-a494-a5ecf5b88cec-image-import-ca\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.588899 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pgqsl\" (UniqueName: \"kubernetes.io/projected/b5fca2ad-c70a-43e2-aff1-f180617a94f8-kube-api-access-pgqsl\") pod \"dns-operator-744455d44c-9jjwn\" (UID: \"b5fca2ad-c70a-43e2-aff1-f180617a94f8\") " pod="openshift-dns-operator/dns-operator-744455d44c-9jjwn" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.588920 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a4ef12d4-a717-4afa-b3c7-179ea0f7f16e-serving-cert\") pod \"console-operator-58897d9998-rt5vd\" (UID: \"a4ef12d4-a717-4afa-b3c7-179ea0f7f16e\") " pod="openshift-console-operator/console-operator-58897d9998-rt5vd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.589031 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b3efa80-8370-4353-b688-4701c4deec8f-config\") pod \"machine-api-operator-5694c8668f-mrsjk\" (UID: \"9b3efa80-8370-4353-b688-4701c4deec8f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mrsjk" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.589063 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2gb75\" (UniqueName: \"kubernetes.io/projected/038209e7-3fdb-490b-976e-f1791bff29d8-kube-api-access-2gb75\") pod \"machine-config-controller-84d6567774-cwz9p\" (UID: \"038209e7-3fdb-490b-976e-f1791bff29d8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cwz9p" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.589541 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-csl6r"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.589574 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e20a512-8f34-4d21-be45-e749f7a6cc87-trusted-ca-bundle\") pod \"console-f9d7485db-csl6r\" (UID: \"5e20a512-8f34-4d21-be45-e749f7a6cc87\") " pod="openshift-console/console-f9d7485db-csl6r" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.589601 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/45ebd02f-dc19-48a5-8309-bdbf11a8a7a4-etcd-ca\") pod \"etcd-operator-b45778765-7sxxj\" (UID: \"45ebd02f-dc19-48a5-8309-bdbf11a8a7a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7sxxj" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.589629 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1c77fe9a-d2e1-465b-8600-9f7f2d1fa206-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-92sz4\" (UID: \"1c77fe9a-d2e1-465b-8600-9f7f2d1fa206\") " pod="openshift-controller-manager/controller-manager-879f6c89f-92sz4" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.589666 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.589686 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0dae89fc-445f-4531-8928-29b334636522-audit-policies\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.589704 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.589726 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f4cc87b-f2fd-4c61-82fb-f362b81445d7-config\") pod \"openshift-apiserver-operator-796bbdcf4f-bjc5w\" (UID: \"8f4cc87b-f2fd-4c61-82fb-f362b81445d7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bjc5w" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.589746 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.589801 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d89bf2d0-8109-40be-b1ce-26ee92a58d81-serving-cert\") pod \"openshift-config-operator-7777fb866f-ws6r9\" (UID: \"d89bf2d0-8109-40be-b1ce-26ee92a58d81\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ws6r9" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.589830 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7fa0bf77-28c4-492d-a4d4-b0a0ec401cf0-metrics-tls\") pod \"ingress-operator-5b745b69d9-n6qrf\" (UID: \"7fa0bf77-28c4-492d-a4d4-b0a0ec401cf0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n6qrf" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.589848 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4ef12d4-a717-4afa-b3c7-179ea0f7f16e-config\") pod \"console-operator-58897d9998-rt5vd\" (UID: \"a4ef12d4-a717-4afa-b3c7-179ea0f7f16e\") " pod="openshift-console-operator/console-operator-58897d9998-rt5vd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.589867 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45ebd02f-dc19-48a5-8309-bdbf11a8a7a4-config\") pod \"etcd-operator-b45778765-7sxxj\" (UID: \"45ebd02f-dc19-48a5-8309-bdbf11a8a7a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7sxxj" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.589889 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.589919 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9b3efa80-8370-4353-b688-4701c4deec8f-images\") pod \"machine-api-operator-5694c8668f-mrsjk\" (UID: \"9b3efa80-8370-4353-b688-4701c4deec8f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mrsjk" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.590029 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/454804e7-0d4d-4536-8544-87a5b6368f70-audit-policies\") pod \"apiserver-7bbb656c7d-nmzrx\" (UID: \"454804e7-0d4d-4536-8544-87a5b6368f70\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.590061 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/454804e7-0d4d-4536-8544-87a5b6368f70-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-nmzrx\" (UID: \"454804e7-0d4d-4536-8544-87a5b6368f70\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.590098 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wpmjz\" (UniqueName: \"kubernetes.io/projected/943c9c2f-24f5-408c-8e8d-c344253fc8f6-kube-api-access-wpmjz\") pod \"machine-approver-56656f9798-vwp4c\" (UID: \"943c9c2f-24f5-408c-8e8d-c344253fc8f6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwp4c" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.590125 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/efe9c94d-8f97-4274-a494-a5ecf5b88cec-etcd-serving-ca\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.590145 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xtvtk\" (UniqueName: \"kubernetes.io/projected/0dae89fc-445f-4531-8928-29b334636522-kube-api-access-xtvtk\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.590165 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b173cdfc-6a31-4d2e-93cd-a04eb1c68877-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rw8ng\" (UID: \"b173cdfc-6a31-4d2e-93cd-a04eb1c68877\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rw8ng" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.590184 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7fa0bf77-28c4-492d-a4d4-b0a0ec401cf0-bound-sa-token\") pod \"ingress-operator-5b745b69d9-n6qrf\" (UID: \"7fa0bf77-28c4-492d-a4d4-b0a0ec401cf0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n6qrf" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.590221 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.590256 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c77fe9a-d2e1-465b-8600-9f7f2d1fa206-serving-cert\") pod \"controller-manager-879f6c89f-92sz4\" (UID: \"1c77fe9a-d2e1-465b-8600-9f7f2d1fa206\") " pod="openshift-controller-manager/controller-manager-879f6c89f-92sz4" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.590276 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qwv5t\" (UniqueName: \"kubernetes.io/projected/89f0e615-1fde-483d-9b38-83287f104360-kube-api-access-qwv5t\") pod \"olm-operator-6b444d44fb-bxjls\" (UID: \"89f0e615-1fde-483d-9b38-83287f104360\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bxjls" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.590306 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/efe9c94d-8f97-4274-a494-a5ecf5b88cec-etcd-client\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.590329 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6j54n\" (UniqueName: \"kubernetes.io/projected/7b200d8c-28e7-494b-99ec-893f1d9ba3ca-kube-api-access-6j54n\") pod \"machine-config-operator-74547568cd-cp2n5\" (UID: \"7b200d8c-28e7-494b-99ec-893f1d9ba3ca\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cp2n5" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.590375 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/efe9c94d-8f97-4274-a494-a5ecf5b88cec-node-pullsecrets\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.590392 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5e20a512-8f34-4d21-be45-e749f7a6cc87-console-config\") pod \"console-f9d7485db-csl6r\" (UID: \"5e20a512-8f34-4d21-be45-e749f7a6cc87\") " pod="openshift-console/console-f9d7485db-csl6r" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.590423 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bc0bb1d-5b22-4e74-8aa8-6554b77a2b73-config\") pod \"authentication-operator-69f744f599-qn4g8\" (UID: \"3bc0bb1d-5b22-4e74-8aa8-6554b77a2b73\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qn4g8" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.590443 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/efe9c94d-8f97-4274-a494-a5ecf5b88cec-audit\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.590549 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/efe9c94d-8f97-4274-a494-a5ecf5b88cec-encryption-config\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.590573 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/efe9c94d-8f97-4274-a494-a5ecf5b88cec-audit-dir\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.590589 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5449c948-487e-426d-b151-31e8079576f3-client-ca\") pod \"route-controller-manager-6576b87f9c-wtn79\" (UID: \"5449c948-487e-426d-b151-31e8079576f3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wtn79" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.590608 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7fa0bf77-28c4-492d-a4d4-b0a0ec401cf0-trusted-ca\") pod \"ingress-operator-5b745b69d9-n6qrf\" (UID: \"7fa0bf77-28c4-492d-a4d4-b0a0ec401cf0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n6qrf" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.590628 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v7dwc\" (UniqueName: \"kubernetes.io/projected/c3a20ddb-cfaa-46ac-8aed-4db9de3e32b6-kube-api-access-v7dwc\") pod \"control-plane-machine-set-operator-78cbb6b69f-rv6n2\" (UID: \"c3a20ddb-cfaa-46ac-8aed-4db9de3e32b6\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rv6n2" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.590646 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/038209e7-3fdb-490b-976e-f1791bff29d8-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-cwz9p\" (UID: \"038209e7-3fdb-490b-976e-f1791bff29d8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cwz9p" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.590675 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wf8bd\" (UniqueName: \"kubernetes.io/projected/d89bf2d0-8109-40be-b1ce-26ee92a58d81-kube-api-access-wf8bd\") pod \"openshift-config-operator-7777fb866f-ws6r9\" (UID: \"d89bf2d0-8109-40be-b1ce-26ee92a58d81\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ws6r9" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.590718 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3bc0bb1d-5b22-4e74-8aa8-6554b77a2b73-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-qn4g8\" (UID: \"3bc0bb1d-5b22-4e74-8aa8-6554b77a2b73\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qn4g8" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.590740 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.590767 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3bc0bb1d-5b22-4e74-8aa8-6554b77a2b73-service-ca-bundle\") pod \"authentication-operator-69f744f599-qn4g8\" (UID: \"3bc0bb1d-5b22-4e74-8aa8-6554b77a2b73\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qn4g8" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.590801 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/45ebd02f-dc19-48a5-8309-bdbf11a8a7a4-etcd-client\") pod \"etcd-operator-b45778765-7sxxj\" (UID: \"45ebd02f-dc19-48a5-8309-bdbf11a8a7a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7sxxj" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.590817 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.590859 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1c77fe9a-d2e1-465b-8600-9f7f2d1fa206-client-ca\") pod \"controller-manager-879f6c89f-92sz4\" (UID: \"1c77fe9a-d2e1-465b-8600-9f7f2d1fa206\") " pod="openshift-controller-manager/controller-manager-879f6c89f-92sz4" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.590875 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b173cdfc-6a31-4d2e-93cd-a04eb1c68877-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rw8ng\" (UID: \"b173cdfc-6a31-4d2e-93cd-a04eb1c68877\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rw8ng" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.590931 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/89f0e615-1fde-483d-9b38-83287f104360-srv-cert\") pod \"olm-operator-6b444d44fb-bxjls\" (UID: \"89f0e615-1fde-483d-9b38-83287f104360\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bxjls" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.590960 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b5fca2ad-c70a-43e2-aff1-f180617a94f8-metrics-tls\") pod \"dns-operator-744455d44c-9jjwn\" (UID: \"b5fca2ad-c70a-43e2-aff1-f180617a94f8\") " pod="openshift-dns-operator/dns-operator-744455d44c-9jjwn" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.591081 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sspzl\" (UniqueName: \"kubernetes.io/projected/14860b75-6fa1-43b5-ba20-d731531dba3b-kube-api-access-sspzl\") pod \"cluster-samples-operator-665b6dd947-ws79r\" (UID: \"14860b75-6fa1-43b5-ba20-d731531dba3b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ws79r" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.591102 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5e20a512-8f34-4d21-be45-e749f7a6cc87-oauth-serving-cert\") pod \"console-f9d7485db-csl6r\" (UID: \"5e20a512-8f34-4d21-be45-e749f7a6cc87\") " pod="openshift-console/console-f9d7485db-csl6r" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.591120 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45ebd02f-dc19-48a5-8309-bdbf11a8a7a4-serving-cert\") pod \"etcd-operator-b45778765-7sxxj\" (UID: \"45ebd02f-dc19-48a5-8309-bdbf11a8a7a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7sxxj" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.591149 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5e20a512-8f34-4d21-be45-e749f7a6cc87-console-serving-cert\") pod \"console-f9d7485db-csl6r\" (UID: \"5e20a512-8f34-4d21-be45-e749f7a6cc87\") " pod="openshift-console/console-f9d7485db-csl6r" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.591174 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zwlq2\" (UniqueName: \"kubernetes.io/projected/5e20a512-8f34-4d21-be45-e749f7a6cc87-kube-api-access-zwlq2\") pod \"console-f9d7485db-csl6r\" (UID: \"5e20a512-8f34-4d21-be45-e749f7a6cc87\") " pod="openshift-console/console-f9d7485db-csl6r" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.591190 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5e20a512-8f34-4d21-be45-e749f7a6cc87-console-oauth-config\") pod \"console-f9d7485db-csl6r\" (UID: \"5e20a512-8f34-4d21-be45-e749f7a6cc87\") " pod="openshift-console/console-f9d7485db-csl6r" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.591206 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/454804e7-0d4d-4536-8544-87a5b6368f70-etcd-client\") pod \"apiserver-7bbb656c7d-nmzrx\" (UID: \"454804e7-0d4d-4536-8544-87a5b6368f70\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.591230 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7b200d8c-28e7-494b-99ec-893f1d9ba3ca-proxy-tls\") pod \"machine-config-operator-74547568cd-cp2n5\" (UID: \"7b200d8c-28e7-494b-99ec-893f1d9ba3ca\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cp2n5" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.591256 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tdtmh\" (UniqueName: \"kubernetes.io/projected/d6b9a450-fd86-47a3-9c0e-f61aa7a24377-kube-api-access-tdtmh\") pod \"downloads-7954f5f757-dtj88\" (UID: \"d6b9a450-fd86-47a3-9c0e-f61aa7a24377\") " pod="openshift-console/downloads-7954f5f757-dtj88" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.591390 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/454804e7-0d4d-4536-8544-87a5b6368f70-encryption-config\") pod \"apiserver-7bbb656c7d-nmzrx\" (UID: \"454804e7-0d4d-4536-8544-87a5b6368f70\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.591410 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efe9c94d-8f97-4274-a494-a5ecf5b88cec-config\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.591446 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/454804e7-0d4d-4536-8544-87a5b6368f70-audit-dir\") pod \"apiserver-7bbb656c7d-nmzrx\" (UID: \"454804e7-0d4d-4536-8544-87a5b6368f70\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.591468 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5449c948-487e-426d-b151-31e8079576f3-config\") pod \"route-controller-manager-6576b87f9c-wtn79\" (UID: \"5449c948-487e-426d-b151-31e8079576f3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wtn79" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.591487 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/d89bf2d0-8109-40be-b1ce-26ee92a58d81-available-featuregates\") pod \"openshift-config-operator-7777fb866f-ws6r9\" (UID: \"d89bf2d0-8109-40be-b1ce-26ee92a58d81\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ws6r9" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.591528 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hk288\" (UniqueName: \"kubernetes.io/projected/efe9c94d-8f97-4274-a494-a5ecf5b88cec-kube-api-access-hk288\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.591551 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7b200d8c-28e7-494b-99ec-893f1d9ba3ca-images\") pod \"machine-config-operator-74547568cd-cp2n5\" (UID: \"7b200d8c-28e7-494b-99ec-893f1d9ba3ca\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cp2n5" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.591611 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/092da80d-ee9c-4e88-b3e2-58bb3dd77f9b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-tr7z7\" (UID: \"092da80d-ee9c-4e88-b3e2-58bb3dd77f9b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tr7z7" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.591631 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/c3a20ddb-cfaa-46ac-8aed-4db9de3e32b6-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-rv6n2\" (UID: \"c3a20ddb-cfaa-46ac-8aed-4db9de3e32b6\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rv6n2" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.591668 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/49347426-bca5-4fde-8b87-21dc8fc486d8-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-kmpsc\" (UID: \"49347426-bca5-4fde-8b87-21dc8fc486d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kmpsc" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.591688 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/038209e7-3fdb-490b-976e-f1791bff29d8-proxy-tls\") pod \"machine-config-controller-84d6567774-cwz9p\" (UID: \"038209e7-3fdb-490b-976e-f1791bff29d8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cwz9p" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.591703 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f4cc87b-f2fd-4c61-82fb-f362b81445d7-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-bjc5w\" (UID: \"8f4cc87b-f2fd-4c61-82fb-f362b81445d7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bjc5w" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.591744 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8vnp\" (UniqueName: \"kubernetes.io/projected/1c77fe9a-d2e1-465b-8600-9f7f2d1fa206-kube-api-access-j8vnp\") pod \"controller-manager-879f6c89f-92sz4\" (UID: \"1c77fe9a-d2e1-465b-8600-9f7f2d1fa206\") " pod="openshift-controller-manager/controller-manager-879f6c89f-92sz4" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.591765 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qvzlg\" (UniqueName: \"kubernetes.io/projected/5449c948-487e-426d-b151-31e8079576f3-kube-api-access-qvzlg\") pod \"route-controller-manager-6576b87f9c-wtn79\" (UID: \"5449c948-487e-426d-b151-31e8079576f3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wtn79" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.591788 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/efe9c94d-8f97-4274-a494-a5ecf5b88cec-serving-cert\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.591823 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7b200d8c-28e7-494b-99ec-893f1d9ba3ca-auth-proxy-config\") pod \"machine-config-operator-74547568cd-cp2n5\" (UID: \"7b200d8c-28e7-494b-99ec-893f1d9ba3ca\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cp2n5" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.591844 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/092da80d-ee9c-4e88-b3e2-58bb3dd77f9b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-tr7z7\" (UID: \"092da80d-ee9c-4e88-b3e2-58bb3dd77f9b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tr7z7" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.591870 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-shzqz\" (UniqueName: \"kubernetes.io/projected/7fa0bf77-28c4-492d-a4d4-b0a0ec401cf0-kube-api-access-shzqz\") pod \"ingress-operator-5b745b69d9-n6qrf\" (UID: \"7fa0bf77-28c4-492d-a4d4-b0a0ec401cf0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n6qrf" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.591929 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac8fc0d4-aa55-4974-a506-7ccd2713f31a-config\") pod \"kube-apiserver-operator-766d6c64bb-gwrxz\" (UID: \"ac8fc0d4-aa55-4974-a506-7ccd2713f31a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gwrxz" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.591948 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.591985 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/454804e7-0d4d-4536-8544-87a5b6368f70-serving-cert\") pod \"apiserver-7bbb656c7d-nmzrx\" (UID: \"454804e7-0d4d-4536-8544-87a5b6368f70\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.592006 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/943c9c2f-24f5-408c-8e8d-c344253fc8f6-config\") pod \"machine-approver-56656f9798-vwp4c\" (UID: \"943c9c2f-24f5-408c-8e8d-c344253fc8f6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwp4c" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.592025 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac8fc0d4-aa55-4974-a506-7ccd2713f31a-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-gwrxz\" (UID: \"ac8fc0d4-aa55-4974-a506-7ccd2713f31a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gwrxz" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.592110 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0dae89fc-445f-4531-8928-29b334636522-audit-dir\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.593266 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/454804e7-0d4d-4536-8544-87a5b6368f70-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-nmzrx\" (UID: \"454804e7-0d4d-4536-8544-87a5b6368f70\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.593535 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qg482\" (UniqueName: \"kubernetes.io/projected/454804e7-0d4d-4536-8544-87a5b6368f70-kube-api-access-qg482\") pod \"apiserver-7bbb656c7d-nmzrx\" (UID: \"454804e7-0d4d-4536-8544-87a5b6368f70\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.593558 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1c77fe9a-d2e1-465b-8600-9f7f2d1fa206-client-ca\") pod \"controller-manager-879f6c89f-92sz4\" (UID: \"1c77fe9a-d2e1-465b-8600-9f7f2d1fa206\") " pod="openshift-controller-manager/controller-manager-879f6c89f-92sz4" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.593572 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3883197f-52dd-465a-82bb-c4dd32b3d090-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rmkqj\" (UID: \"3883197f-52dd-465a-82bb-c4dd32b3d090\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rmkqj" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.595592 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-dnnwg"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.595616 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/efe9c94d-8f97-4274-a494-a5ecf5b88cec-image-import-ca\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.596007 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/efe9c94d-8f97-4274-a494-a5ecf5b88cec-node-pullsecrets\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.596100 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.596151 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ac8fc0d4-aa55-4974-a506-7ccd2713f31a-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-gwrxz\" (UID: \"ac8fc0d4-aa55-4974-a506-7ccd2713f31a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gwrxz" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.601822 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-dnnwg" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.602096 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/efe9c94d-8f97-4274-a494-a5ecf5b88cec-audit-dir\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.602539 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/efe9c94d-8f97-4274-a494-a5ecf5b88cec-config\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.602599 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9b3efa80-8370-4353-b688-4701c4deec8f-config\") pod \"machine-api-operator-5694c8668f-mrsjk\" (UID: \"9b3efa80-8370-4353-b688-4701c4deec8f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mrsjk" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.602676 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1c77fe9a-d2e1-465b-8600-9f7f2d1fa206-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-92sz4\" (UID: \"1c77fe9a-d2e1-465b-8600-9f7f2d1fa206\") " pod="openshift-controller-manager/controller-manager-879f6c89f-92sz4" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.602979 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wk44b\" (UniqueName: \"kubernetes.io/projected/49347426-bca5-4fde-8b87-21dc8fc486d8-kube-api-access-wk44b\") pod \"cluster-image-registry-operator-dc59b4c8b-kmpsc\" (UID: \"49347426-bca5-4fde-8b87-21dc8fc486d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kmpsc" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.603010 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/454804e7-0d4d-4536-8544-87a5b6368f70-audit-dir\") pod \"apiserver-7bbb656c7d-nmzrx\" (UID: \"454804e7-0d4d-4536-8544-87a5b6368f70\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.603015 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/89f0e615-1fde-483d-9b38-83287f104360-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bxjls\" (UID: \"89f0e615-1fde-483d-9b38-83287f104360\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bxjls" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.602991 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/efe9c94d-8f97-4274-a494-a5ecf5b88cec-etcd-serving-ca\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.603035 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6sjfd"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.603430 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/454804e7-0d4d-4536-8544-87a5b6368f70-audit-policies\") pod \"apiserver-7bbb656c7d-nmzrx\" (UID: \"454804e7-0d4d-4536-8544-87a5b6368f70\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.603490 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/9b3efa80-8370-4353-b688-4701c4deec8f-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-mrsjk\" (UID: \"9b3efa80-8370-4353-b688-4701c4deec8f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mrsjk" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.603528 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3883197f-52dd-465a-82bb-c4dd32b3d090-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rmkqj\" (UID: \"3883197f-52dd-465a-82bb-c4dd32b3d090\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rmkqj" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.603553 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xs6pj\" (UniqueName: \"kubernetes.io/projected/45ebd02f-dc19-48a5-8309-bdbf11a8a7a4-kube-api-access-xs6pj\") pod \"etcd-operator-b45778765-7sxxj\" (UID: \"45ebd02f-dc19-48a5-8309-bdbf11a8a7a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7sxxj" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.603571 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bmclp\" (UniqueName: \"kubernetes.io/projected/8f4cc87b-f2fd-4c61-82fb-f362b81445d7-kube-api-access-bmclp\") pod \"openshift-apiserver-operator-796bbdcf4f-bjc5w\" (UID: \"8f4cc87b-f2fd-4c61-82fb-f362b81445d7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bjc5w" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.603589 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/454804e7-0d4d-4536-8544-87a5b6368f70-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-nmzrx\" (UID: \"454804e7-0d4d-4536-8544-87a5b6368f70\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.603593 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/efe9c94d-8f97-4274-a494-a5ecf5b88cec-trusted-ca-bundle\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.603710 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/49347426-bca5-4fde-8b87-21dc8fc486d8-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-kmpsc\" (UID: \"49347426-bca5-4fde-8b87-21dc8fc486d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kmpsc" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.603733 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/14860b75-6fa1-43b5-ba20-d731531dba3b-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-ws79r\" (UID: \"14860b75-6fa1-43b5-ba20-d731531dba3b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ws79r" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.603755 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/45ebd02f-dc19-48a5-8309-bdbf11a8a7a4-etcd-service-ca\") pod \"etcd-operator-b45778765-7sxxj\" (UID: \"45ebd02f-dc19-48a5-8309-bdbf11a8a7a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7sxxj" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.603814 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.603842 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/943c9c2f-24f5-408c-8e8d-c344253fc8f6-machine-approver-tls\") pod \"machine-approver-56656f9798-vwp4c\" (UID: \"943c9c2f-24f5-408c-8e8d-c344253fc8f6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwp4c" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.603861 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5e20a512-8f34-4d21-be45-e749f7a6cc87-service-ca\") pod \"console-f9d7485db-csl6r\" (UID: \"5e20a512-8f34-4d21-be45-e749f7a6cc87\") " pod="openshift-console/console-f9d7485db-csl6r" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.603878 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6vbsf\" (UniqueName: \"kubernetes.io/projected/81c6feb2-e6ea-43f0-834d-499bea65b4aa-kube-api-access-6vbsf\") pod \"migrator-59844c95c7-gbqfq\" (UID: \"81c6feb2-e6ea-43f0-834d-499bea65b4aa\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gbqfq" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.603896 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/943c9c2f-24f5-408c-8e8d-c344253fc8f6-auth-proxy-config\") pod \"machine-approver-56656f9798-vwp4c\" (UID: \"943c9c2f-24f5-408c-8e8d-c344253fc8f6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwp4c" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.603955 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b173cdfc-6a31-4d2e-93cd-a04eb1c68877-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rw8ng\" (UID: \"b173cdfc-6a31-4d2e-93cd-a04eb1c68877\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rw8ng" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.603975 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xlsbr\" (UniqueName: \"kubernetes.io/projected/a4ef12d4-a717-4afa-b3c7-179ea0f7f16e-kube-api-access-xlsbr\") pod \"console-operator-58897d9998-rt5vd\" (UID: \"a4ef12d4-a717-4afa-b3c7-179ea0f7f16e\") " pod="openshift-console-operator/console-operator-58897d9998-rt5vd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.604497 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/efe9c94d-8f97-4274-a494-a5ecf5b88cec-trusted-ca-bundle\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.604504 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.604537 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t6s92\" (UniqueName: \"kubernetes.io/projected/3883197f-52dd-465a-82bb-c4dd32b3d090-kube-api-access-t6s92\") pod \"kube-storage-version-migrator-operator-b67b599dd-rmkqj\" (UID: \"3883197f-52dd-465a-82bb-c4dd32b3d090\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rmkqj" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.604656 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5449c948-487e-426d-b151-31e8079576f3-serving-cert\") pod \"route-controller-manager-6576b87f9c-wtn79\" (UID: \"5449c948-487e-426d-b151-31e8079576f3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wtn79" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.604676 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a4ef12d4-a717-4afa-b3c7-179ea0f7f16e-trusted-ca\") pod \"console-operator-58897d9998-rt5vd\" (UID: \"a4ef12d4-a717-4afa-b3c7-179ea0f7f16e\") " pod="openshift-console-operator/console-operator-58897d9998-rt5vd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.604692 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3bc0bb1d-5b22-4e74-8aa8-6554b77a2b73-serving-cert\") pod \"authentication-operator-69f744f599-qn4g8\" (UID: \"3bc0bb1d-5b22-4e74-8aa8-6554b77a2b73\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qn4g8" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.604816 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c77fe9a-d2e1-465b-8600-9f7f2d1fa206-config\") pod \"controller-manager-879f6c89f-92sz4\" (UID: \"1c77fe9a-d2e1-465b-8600-9f7f2d1fa206\") " pod="openshift-controller-manager/controller-manager-879f6c89f-92sz4" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.605403 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-cwz9p"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.605465 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.605780 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/9b3efa80-8370-4353-b688-4701c4deec8f-images\") pod \"machine-api-operator-5694c8668f-mrsjk\" (UID: \"9b3efa80-8370-4353-b688-4701c4deec8f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mrsjk" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.606446 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7sxxj"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.606692 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/efe9c94d-8f97-4274-a494-a5ecf5b88cec-audit\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.606994 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/efe9c94d-8f97-4274-a494-a5ecf5b88cec-serving-cert\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.607195 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/454804e7-0d4d-4536-8544-87a5b6368f70-serving-cert\") pod \"apiserver-7bbb656c7d-nmzrx\" (UID: \"454804e7-0d4d-4536-8544-87a5b6368f70\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.607505 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-9jjwn"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.608197 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/9b3efa80-8370-4353-b688-4701c4deec8f-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-mrsjk\" (UID: \"9b3efa80-8370-4353-b688-4701c4deec8f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mrsjk" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.608347 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/454804e7-0d4d-4536-8544-87a5b6368f70-encryption-config\") pod \"apiserver-7bbb656c7d-nmzrx\" (UID: \"454804e7-0d4d-4536-8544-87a5b6368f70\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.608659 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.609367 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/efe9c94d-8f97-4274-a494-a5ecf5b88cec-encryption-config\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.610102 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-ws6r9"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.611010 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/efe9c94d-8f97-4274-a494-a5ecf5b88cec-etcd-client\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.611062 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wtn79"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.611114 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c77fe9a-d2e1-465b-8600-9f7f2d1fa206-serving-cert\") pod \"controller-manager-879f6c89f-92sz4\" (UID: \"1c77fe9a-d2e1-465b-8600-9f7f2d1fa206\") " pod="openshift-controller-manager/controller-manager-879f6c89f-92sz4" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.612518 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bjc5w"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.613529 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnjpj"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.614693 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-gbqfq"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.615698 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rv6n2"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.616668 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tr7z7"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.617841 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7nk8v"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.619529 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-fsc8x"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.620144 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-fsc8x" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.620426 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/454804e7-0d4d-4536-8544-87a5b6368f70-etcd-client\") pod \"apiserver-7bbb656c7d-nmzrx\" (UID: \"454804e7-0d4d-4536-8544-87a5b6368f70\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.621053 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-d8x8n"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.622155 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-d8x8n" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.623012 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-qn4g8"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.623773 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.625297 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rmkqj"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.628632 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-n6qrf"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.629982 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v7xvv"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.633620 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-49rxs"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.634983 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335905-kdpld"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.638191 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rw8ng"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.641999 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-4xkx2"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.642034 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-g4q8m"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.642047 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-dnnwg"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.645012 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-cp2n5"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.645039 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-92sz4"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.647306 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-d8x8n"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.647544 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mhnj2"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.649579 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-9c49l"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.650288 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hk2cx"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.650665 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.651744 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-jg6t8"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.652793 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-jg6t8"] Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.652958 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-jg6t8" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.665570 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.691482 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.704349 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705474 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b173cdfc-6a31-4d2e-93cd-a04eb1c68877-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rw8ng\" (UID: \"b173cdfc-6a31-4d2e-93cd-a04eb1c68877\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rw8ng" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705500 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xlsbr\" (UniqueName: \"kubernetes.io/projected/a4ef12d4-a717-4afa-b3c7-179ea0f7f16e-kube-api-access-xlsbr\") pod \"console-operator-58897d9998-rt5vd\" (UID: \"a4ef12d4-a717-4afa-b3c7-179ea0f7f16e\") " pod="openshift-console-operator/console-operator-58897d9998-rt5vd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705524 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705543 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t6s92\" (UniqueName: \"kubernetes.io/projected/3883197f-52dd-465a-82bb-c4dd32b3d090-kube-api-access-t6s92\") pod \"kube-storage-version-migrator-operator-b67b599dd-rmkqj\" (UID: \"3883197f-52dd-465a-82bb-c4dd32b3d090\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rmkqj" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705559 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5449c948-487e-426d-b151-31e8079576f3-serving-cert\") pod \"route-controller-manager-6576b87f9c-wtn79\" (UID: \"5449c948-487e-426d-b151-31e8079576f3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wtn79" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705574 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a4ef12d4-a717-4afa-b3c7-179ea0f7f16e-trusted-ca\") pod \"console-operator-58897d9998-rt5vd\" (UID: \"a4ef12d4-a717-4afa-b3c7-179ea0f7f16e\") " pod="openshift-console-operator/console-operator-58897d9998-rt5vd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705587 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3bc0bb1d-5b22-4e74-8aa8-6554b77a2b73-serving-cert\") pod \"authentication-operator-69f744f599-qn4g8\" (UID: \"3bc0bb1d-5b22-4e74-8aa8-6554b77a2b73\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qn4g8" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705603 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kp59m\" (UniqueName: \"kubernetes.io/projected/092da80d-ee9c-4e88-b3e2-58bb3dd77f9b-kube-api-access-kp59m\") pod \"openshift-controller-manager-operator-756b6f6bc6-tr7z7\" (UID: \"092da80d-ee9c-4e88-b3e2-58bb3dd77f9b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tr7z7" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705620 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/23902dde-0c8d-4d54-99f4-0ee4e19a1a50-certs\") pod \"machine-config-server-fsc8x\" (UID: \"23902dde-0c8d-4d54-99f4-0ee4e19a1a50\") " pod="openshift-machine-config-operator/machine-config-server-fsc8x" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705646 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/49347426-bca5-4fde-8b87-21dc8fc486d8-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-kmpsc\" (UID: \"49347426-bca5-4fde-8b87-21dc8fc486d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kmpsc" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705663 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rwhh8\" (UniqueName: \"kubernetes.io/projected/3bc0bb1d-5b22-4e74-8aa8-6554b77a2b73-kube-api-access-rwhh8\") pod \"authentication-operator-69f744f599-qn4g8\" (UID: \"3bc0bb1d-5b22-4e74-8aa8-6554b77a2b73\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qn4g8" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705677 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pgqsl\" (UniqueName: \"kubernetes.io/projected/b5fca2ad-c70a-43e2-aff1-f180617a94f8-kube-api-access-pgqsl\") pod \"dns-operator-744455d44c-9jjwn\" (UID: \"b5fca2ad-c70a-43e2-aff1-f180617a94f8\") " pod="openshift-dns-operator/dns-operator-744455d44c-9jjwn" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705691 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a4ef12d4-a717-4afa-b3c7-179ea0f7f16e-serving-cert\") pod \"console-operator-58897d9998-rt5vd\" (UID: \"a4ef12d4-a717-4afa-b3c7-179ea0f7f16e\") " pod="openshift-console-operator/console-operator-58897d9998-rt5vd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705705 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2gb75\" (UniqueName: \"kubernetes.io/projected/038209e7-3fdb-490b-976e-f1791bff29d8-kube-api-access-2gb75\") pod \"machine-config-controller-84d6567774-cwz9p\" (UID: \"038209e7-3fdb-490b-976e-f1791bff29d8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cwz9p" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705719 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e20a512-8f34-4d21-be45-e749f7a6cc87-trusted-ca-bundle\") pod \"console-f9d7485db-csl6r\" (UID: \"5e20a512-8f34-4d21-be45-e749f7a6cc87\") " pod="openshift-console/console-f9d7485db-csl6r" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705733 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/45ebd02f-dc19-48a5-8309-bdbf11a8a7a4-etcd-ca\") pod \"etcd-operator-b45778765-7sxxj\" (UID: \"45ebd02f-dc19-48a5-8309-bdbf11a8a7a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7sxxj" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705747 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705761 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0dae89fc-445f-4531-8928-29b334636522-audit-policies\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705776 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705794 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f4cc87b-f2fd-4c61-82fb-f362b81445d7-config\") pod \"openshift-apiserver-operator-796bbdcf4f-bjc5w\" (UID: \"8f4cc87b-f2fd-4c61-82fb-f362b81445d7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bjc5w" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705810 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705824 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d89bf2d0-8109-40be-b1ce-26ee92a58d81-serving-cert\") pod \"openshift-config-operator-7777fb866f-ws6r9\" (UID: \"d89bf2d0-8109-40be-b1ce-26ee92a58d81\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ws6r9" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705840 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7fa0bf77-28c4-492d-a4d4-b0a0ec401cf0-metrics-tls\") pod \"ingress-operator-5b745b69d9-n6qrf\" (UID: \"7fa0bf77-28c4-492d-a4d4-b0a0ec401cf0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n6qrf" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705855 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4ef12d4-a717-4afa-b3c7-179ea0f7f16e-config\") pod \"console-operator-58897d9998-rt5vd\" (UID: \"a4ef12d4-a717-4afa-b3c7-179ea0f7f16e\") " pod="openshift-console-operator/console-operator-58897d9998-rt5vd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705869 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45ebd02f-dc19-48a5-8309-bdbf11a8a7a4-config\") pod \"etcd-operator-b45778765-7sxxj\" (UID: \"45ebd02f-dc19-48a5-8309-bdbf11a8a7a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7sxxj" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705885 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705901 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/0761866e-d6ab-40c9-a776-8c7fb2683fad-default-certificate\") pod \"router-default-5444994796-2fvpq\" (UID: \"0761866e-d6ab-40c9-a776-8c7fb2683fad\") " pod="openshift-ingress/router-default-5444994796-2fvpq" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705918 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wpmjz\" (UniqueName: \"kubernetes.io/projected/943c9c2f-24f5-408c-8e8d-c344253fc8f6-kube-api-access-wpmjz\") pod \"machine-approver-56656f9798-vwp4c\" (UID: \"943c9c2f-24f5-408c-8e8d-c344253fc8f6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwp4c" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705933 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fwwhz\" (UniqueName: \"kubernetes.io/projected/c506a84b-4ea6-477e-88de-fceec3a95f31-kube-api-access-fwwhz\") pod \"collect-profiles-29335905-kdpld\" (UID: \"c506a84b-4ea6-477e-88de-fceec3a95f31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-kdpld" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705949 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xtvtk\" (UniqueName: \"kubernetes.io/projected/0dae89fc-445f-4531-8928-29b334636522-kube-api-access-xtvtk\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705963 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c506a84b-4ea6-477e-88de-fceec3a95f31-secret-volume\") pod \"collect-profiles-29335905-kdpld\" (UID: \"c506a84b-4ea6-477e-88de-fceec3a95f31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-kdpld" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705979 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b173cdfc-6a31-4d2e-93cd-a04eb1c68877-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rw8ng\" (UID: \"b173cdfc-6a31-4d2e-93cd-a04eb1c68877\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rw8ng" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.705993 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7fa0bf77-28c4-492d-a4d4-b0a0ec401cf0-bound-sa-token\") pod \"ingress-operator-5b745b69d9-n6qrf\" (UID: \"7fa0bf77-28c4-492d-a4d4-b0a0ec401cf0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n6qrf" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706009 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706024 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qwv5t\" (UniqueName: \"kubernetes.io/projected/89f0e615-1fde-483d-9b38-83287f104360-kube-api-access-qwv5t\") pod \"olm-operator-6b444d44fb-bxjls\" (UID: \"89f0e615-1fde-483d-9b38-83287f104360\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bxjls" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706039 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6j54n\" (UniqueName: \"kubernetes.io/projected/7b200d8c-28e7-494b-99ec-893f1d9ba3ca-kube-api-access-6j54n\") pod \"machine-config-operator-74547568cd-cp2n5\" (UID: \"7b200d8c-28e7-494b-99ec-893f1d9ba3ca\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cp2n5" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706054 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/0761866e-d6ab-40c9-a776-8c7fb2683fad-stats-auth\") pod \"router-default-5444994796-2fvpq\" (UID: \"0761866e-d6ab-40c9-a776-8c7fb2683fad\") " pod="openshift-ingress/router-default-5444994796-2fvpq" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706088 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5e20a512-8f34-4d21-be45-e749f7a6cc87-console-config\") pod \"console-f9d7485db-csl6r\" (UID: \"5e20a512-8f34-4d21-be45-e749f7a6cc87\") " pod="openshift-console/console-f9d7485db-csl6r" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706105 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bc0bb1d-5b22-4e74-8aa8-6554b77a2b73-config\") pod \"authentication-operator-69f744f599-qn4g8\" (UID: \"3bc0bb1d-5b22-4e74-8aa8-6554b77a2b73\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qn4g8" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706120 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v7dwc\" (UniqueName: \"kubernetes.io/projected/c3a20ddb-cfaa-46ac-8aed-4db9de3e32b6-kube-api-access-v7dwc\") pod \"control-plane-machine-set-operator-78cbb6b69f-rv6n2\" (UID: \"c3a20ddb-cfaa-46ac-8aed-4db9de3e32b6\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rv6n2" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706135 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/23902dde-0c8d-4d54-99f4-0ee4e19a1a50-node-bootstrap-token\") pod \"machine-config-server-fsc8x\" (UID: \"23902dde-0c8d-4d54-99f4-0ee4e19a1a50\") " pod="openshift-machine-config-operator/machine-config-server-fsc8x" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706152 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5449c948-487e-426d-b151-31e8079576f3-client-ca\") pod \"route-controller-manager-6576b87f9c-wtn79\" (UID: \"5449c948-487e-426d-b151-31e8079576f3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wtn79" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706169 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7fa0bf77-28c4-492d-a4d4-b0a0ec401cf0-trusted-ca\") pod \"ingress-operator-5b745b69d9-n6qrf\" (UID: \"7fa0bf77-28c4-492d-a4d4-b0a0ec401cf0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n6qrf" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706183 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/038209e7-3fdb-490b-976e-f1791bff29d8-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-cwz9p\" (UID: \"038209e7-3fdb-490b-976e-f1791bff29d8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cwz9p" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706198 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wf8bd\" (UniqueName: \"kubernetes.io/projected/d89bf2d0-8109-40be-b1ce-26ee92a58d81-kube-api-access-wf8bd\") pod \"openshift-config-operator-7777fb866f-ws6r9\" (UID: \"d89bf2d0-8109-40be-b1ce-26ee92a58d81\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ws6r9" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706214 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3bc0bb1d-5b22-4e74-8aa8-6554b77a2b73-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-qn4g8\" (UID: \"3bc0bb1d-5b22-4e74-8aa8-6554b77a2b73\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qn4g8" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706229 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706244 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3bc0bb1d-5b22-4e74-8aa8-6554b77a2b73-service-ca-bundle\") pod \"authentication-operator-69f744f599-qn4g8\" (UID: \"3bc0bb1d-5b22-4e74-8aa8-6554b77a2b73\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qn4g8" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706264 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/45ebd02f-dc19-48a5-8309-bdbf11a8a7a4-etcd-client\") pod \"etcd-operator-b45778765-7sxxj\" (UID: \"45ebd02f-dc19-48a5-8309-bdbf11a8a7a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7sxxj" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706279 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706304 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b5fca2ad-c70a-43e2-aff1-f180617a94f8-metrics-tls\") pod \"dns-operator-744455d44c-9jjwn\" (UID: \"b5fca2ad-c70a-43e2-aff1-f180617a94f8\") " pod="openshift-dns-operator/dns-operator-744455d44c-9jjwn" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706318 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b173cdfc-6a31-4d2e-93cd-a04eb1c68877-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rw8ng\" (UID: \"b173cdfc-6a31-4d2e-93cd-a04eb1c68877\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rw8ng" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706332 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/89f0e615-1fde-483d-9b38-83287f104360-srv-cert\") pod \"olm-operator-6b444d44fb-bxjls\" (UID: \"89f0e615-1fde-483d-9b38-83287f104360\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bxjls" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706348 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sspzl\" (UniqueName: \"kubernetes.io/projected/14860b75-6fa1-43b5-ba20-d731531dba3b-kube-api-access-sspzl\") pod \"cluster-samples-operator-665b6dd947-ws79r\" (UID: \"14860b75-6fa1-43b5-ba20-d731531dba3b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ws79r" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706363 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5e20a512-8f34-4d21-be45-e749f7a6cc87-oauth-serving-cert\") pod \"console-f9d7485db-csl6r\" (UID: \"5e20a512-8f34-4d21-be45-e749f7a6cc87\") " pod="openshift-console/console-f9d7485db-csl6r" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706377 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45ebd02f-dc19-48a5-8309-bdbf11a8a7a4-serving-cert\") pod \"etcd-operator-b45778765-7sxxj\" (UID: \"45ebd02f-dc19-48a5-8309-bdbf11a8a7a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7sxxj" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706405 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5e20a512-8f34-4d21-be45-e749f7a6cc87-console-serving-cert\") pod \"console-f9d7485db-csl6r\" (UID: \"5e20a512-8f34-4d21-be45-e749f7a6cc87\") " pod="openshift-console/console-f9d7485db-csl6r" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706420 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5e20a512-8f34-4d21-be45-e749f7a6cc87-console-oauth-config\") pod \"console-f9d7485db-csl6r\" (UID: \"5e20a512-8f34-4d21-be45-e749f7a6cc87\") " pod="openshift-console/console-f9d7485db-csl6r" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706436 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zwlq2\" (UniqueName: \"kubernetes.io/projected/5e20a512-8f34-4d21-be45-e749f7a6cc87-kube-api-access-zwlq2\") pod \"console-f9d7485db-csl6r\" (UID: \"5e20a512-8f34-4d21-be45-e749f7a6cc87\") " pod="openshift-console/console-f9d7485db-csl6r" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706452 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7b200d8c-28e7-494b-99ec-893f1d9ba3ca-proxy-tls\") pod \"machine-config-operator-74547568cd-cp2n5\" (UID: \"7b200d8c-28e7-494b-99ec-893f1d9ba3ca\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cp2n5" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706468 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tdtmh\" (UniqueName: \"kubernetes.io/projected/d6b9a450-fd86-47a3-9c0e-f61aa7a24377-kube-api-access-tdtmh\") pod \"downloads-7954f5f757-dtj88\" (UID: \"d6b9a450-fd86-47a3-9c0e-f61aa7a24377\") " pod="openshift-console/downloads-7954f5f757-dtj88" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706485 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5449c948-487e-426d-b151-31e8079576f3-config\") pod \"route-controller-manager-6576b87f9c-wtn79\" (UID: \"5449c948-487e-426d-b151-31e8079576f3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wtn79" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706507 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/d89bf2d0-8109-40be-b1ce-26ee92a58d81-available-featuregates\") pod \"openshift-config-operator-7777fb866f-ws6r9\" (UID: \"d89bf2d0-8109-40be-b1ce-26ee92a58d81\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ws6r9" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706528 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7b200d8c-28e7-494b-99ec-893f1d9ba3ca-images\") pod \"machine-config-operator-74547568cd-cp2n5\" (UID: \"7b200d8c-28e7-494b-99ec-893f1d9ba3ca\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cp2n5" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706544 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/092da80d-ee9c-4e88-b3e2-58bb3dd77f9b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-tr7z7\" (UID: \"092da80d-ee9c-4e88-b3e2-58bb3dd77f9b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tr7z7" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706559 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/c3a20ddb-cfaa-46ac-8aed-4db9de3e32b6-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-rv6n2\" (UID: \"c3a20ddb-cfaa-46ac-8aed-4db9de3e32b6\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rv6n2" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706604 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/49347426-bca5-4fde-8b87-21dc8fc486d8-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-kmpsc\" (UID: \"49347426-bca5-4fde-8b87-21dc8fc486d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kmpsc" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706619 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/038209e7-3fdb-490b-976e-f1791bff29d8-proxy-tls\") pod \"machine-config-controller-84d6567774-cwz9p\" (UID: \"038209e7-3fdb-490b-976e-f1791bff29d8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cwz9p" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706633 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f4cc87b-f2fd-4c61-82fb-f362b81445d7-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-bjc5w\" (UID: \"8f4cc87b-f2fd-4c61-82fb-f362b81445d7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bjc5w" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706649 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxdkj\" (UniqueName: \"kubernetes.io/projected/47a99167-adc4-40b4-9925-778c670ad55f-kube-api-access-gxdkj\") pod \"multus-admission-controller-857f4d67dd-g4q8m\" (UID: \"47a99167-adc4-40b4-9925-778c670ad55f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-g4q8m" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706665 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qvzlg\" (UniqueName: \"kubernetes.io/projected/5449c948-487e-426d-b151-31e8079576f3-kube-api-access-qvzlg\") pod \"route-controller-manager-6576b87f9c-wtn79\" (UID: \"5449c948-487e-426d-b151-31e8079576f3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wtn79" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706687 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7b200d8c-28e7-494b-99ec-893f1d9ba3ca-auth-proxy-config\") pod \"machine-config-operator-74547568cd-cp2n5\" (UID: \"7b200d8c-28e7-494b-99ec-893f1d9ba3ca\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cp2n5" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706702 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/092da80d-ee9c-4e88-b3e2-58bb3dd77f9b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-tr7z7\" (UID: \"092da80d-ee9c-4e88-b3e2-58bb3dd77f9b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tr7z7" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706723 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-shzqz\" (UniqueName: \"kubernetes.io/projected/7fa0bf77-28c4-492d-a4d4-b0a0ec401cf0-kube-api-access-shzqz\") pod \"ingress-operator-5b745b69d9-n6qrf\" (UID: \"7fa0bf77-28c4-492d-a4d4-b0a0ec401cf0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n6qrf" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706741 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac8fc0d4-aa55-4974-a506-7ccd2713f31a-config\") pod \"kube-apiserver-operator-766d6c64bb-gwrxz\" (UID: \"ac8fc0d4-aa55-4974-a506-7ccd2713f31a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gwrxz" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706758 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706774 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/943c9c2f-24f5-408c-8e8d-c344253fc8f6-config\") pod \"machine-approver-56656f9798-vwp4c\" (UID: \"943c9c2f-24f5-408c-8e8d-c344253fc8f6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwp4c" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706788 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0dae89fc-445f-4531-8928-29b334636522-audit-dir\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706803 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac8fc0d4-aa55-4974-a506-7ccd2713f31a-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-gwrxz\" (UID: \"ac8fc0d4-aa55-4974-a506-7ccd2713f31a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gwrxz" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706819 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0761866e-d6ab-40c9-a776-8c7fb2683fad-service-ca-bundle\") pod \"router-default-5444994796-2fvpq\" (UID: \"0761866e-d6ab-40c9-a776-8c7fb2683fad\") " pod="openshift-ingress/router-default-5444994796-2fvpq" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706840 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3883197f-52dd-465a-82bb-c4dd32b3d090-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rmkqj\" (UID: \"3883197f-52dd-465a-82bb-c4dd32b3d090\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rmkqj" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706858 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706874 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ac8fc0d4-aa55-4974-a506-7ccd2713f31a-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-gwrxz\" (UID: \"ac8fc0d4-aa55-4974-a506-7ccd2713f31a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gwrxz" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706888 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwxhr\" (UniqueName: \"kubernetes.io/projected/0761866e-d6ab-40c9-a776-8c7fb2683fad-kube-api-access-pwxhr\") pod \"router-default-5444994796-2fvpq\" (UID: \"0761866e-d6ab-40c9-a776-8c7fb2683fad\") " pod="openshift-ingress/router-default-5444994796-2fvpq" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706905 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wk44b\" (UniqueName: \"kubernetes.io/projected/49347426-bca5-4fde-8b87-21dc8fc486d8-kube-api-access-wk44b\") pod \"cluster-image-registry-operator-dc59b4c8b-kmpsc\" (UID: \"49347426-bca5-4fde-8b87-21dc8fc486d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kmpsc" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706921 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/89f0e615-1fde-483d-9b38-83287f104360-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bxjls\" (UID: \"89f0e615-1fde-483d-9b38-83287f104360\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bxjls" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706935 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c506a84b-4ea6-477e-88de-fceec3a95f31-config-volume\") pod \"collect-profiles-29335905-kdpld\" (UID: \"c506a84b-4ea6-477e-88de-fceec3a95f31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-kdpld" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706952 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3883197f-52dd-465a-82bb-c4dd32b3d090-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rmkqj\" (UID: \"3883197f-52dd-465a-82bb-c4dd32b3d090\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rmkqj" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706970 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xs6pj\" (UniqueName: \"kubernetes.io/projected/45ebd02f-dc19-48a5-8309-bdbf11a8a7a4-kube-api-access-xs6pj\") pod \"etcd-operator-b45778765-7sxxj\" (UID: \"45ebd02f-dc19-48a5-8309-bdbf11a8a7a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7sxxj" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.706993 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bmclp\" (UniqueName: \"kubernetes.io/projected/8f4cc87b-f2fd-4c61-82fb-f362b81445d7-kube-api-access-bmclp\") pod \"openshift-apiserver-operator-796bbdcf4f-bjc5w\" (UID: \"8f4cc87b-f2fd-4c61-82fb-f362b81445d7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bjc5w" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.707010 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/49347426-bca5-4fde-8b87-21dc8fc486d8-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-kmpsc\" (UID: \"49347426-bca5-4fde-8b87-21dc8fc486d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kmpsc" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.707018 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/45ebd02f-dc19-48a5-8309-bdbf11a8a7a4-etcd-ca\") pod \"etcd-operator-b45778765-7sxxj\" (UID: \"45ebd02f-dc19-48a5-8309-bdbf11a8a7a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7sxxj" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.707758 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a4ef12d4-a717-4afa-b3c7-179ea0f7f16e-config\") pod \"console-operator-58897d9998-rt5vd\" (UID: \"a4ef12d4-a717-4afa-b3c7-179ea0f7f16e\") " pod="openshift-console-operator/console-operator-58897d9998-rt5vd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.707026 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/14860b75-6fa1-43b5-ba20-d731531dba3b-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-ws79r\" (UID: \"14860b75-6fa1-43b5-ba20-d731531dba3b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ws79r" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.710812 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0761866e-d6ab-40c9-a776-8c7fb2683fad-metrics-certs\") pod \"router-default-5444994796-2fvpq\" (UID: \"0761866e-d6ab-40c9-a776-8c7fb2683fad\") " pod="openshift-ingress/router-default-5444994796-2fvpq" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.710850 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/45ebd02f-dc19-48a5-8309-bdbf11a8a7a4-etcd-service-ca\") pod \"etcd-operator-b45778765-7sxxj\" (UID: \"45ebd02f-dc19-48a5-8309-bdbf11a8a7a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7sxxj" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.710869 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.711093 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/943c9c2f-24f5-408c-8e8d-c344253fc8f6-machine-approver-tls\") pod \"machine-approver-56656f9798-vwp4c\" (UID: \"943c9c2f-24f5-408c-8e8d-c344253fc8f6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwp4c" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.711111 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5e20a512-8f34-4d21-be45-e749f7a6cc87-service-ca\") pod \"console-f9d7485db-csl6r\" (UID: \"5e20a512-8f34-4d21-be45-e749f7a6cc87\") " pod="openshift-console/console-f9d7485db-csl6r" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.711131 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6vbsf\" (UniqueName: \"kubernetes.io/projected/81c6feb2-e6ea-43f0-834d-499bea65b4aa-kube-api-access-6vbsf\") pod \"migrator-59844c95c7-gbqfq\" (UID: \"81c6feb2-e6ea-43f0-834d-499bea65b4aa\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gbqfq" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.711151 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lzw95\" (UniqueName: \"kubernetes.io/projected/23902dde-0c8d-4d54-99f4-0ee4e19a1a50-kube-api-access-lzw95\") pod \"machine-config-server-fsc8x\" (UID: \"23902dde-0c8d-4d54-99f4-0ee4e19a1a50\") " pod="openshift-machine-config-operator/machine-config-server-fsc8x" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.711173 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/943c9c2f-24f5-408c-8e8d-c344253fc8f6-auth-proxy-config\") pod \"machine-approver-56656f9798-vwp4c\" (UID: \"943c9c2f-24f5-408c-8e8d-c344253fc8f6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwp4c" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.711208 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/47a99167-adc4-40b4-9925-778c670ad55f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-g4q8m\" (UID: \"47a99167-adc4-40b4-9925-778c670ad55f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-g4q8m" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.711241 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.708366 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/45ebd02f-dc19-48a5-8309-bdbf11a8a7a4-config\") pod \"etcd-operator-b45778765-7sxxj\" (UID: \"45ebd02f-dc19-48a5-8309-bdbf11a8a7a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7sxxj" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.708845 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5449c948-487e-426d-b151-31e8079576f3-config\") pod \"route-controller-manager-6576b87f9c-wtn79\" (UID: \"5449c948-487e-426d-b151-31e8079576f3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wtn79" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.711669 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.711722 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d89bf2d0-8109-40be-b1ce-26ee92a58d81-serving-cert\") pod \"openshift-config-operator-7777fb866f-ws6r9\" (UID: \"d89bf2d0-8109-40be-b1ce-26ee92a58d81\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ws6r9" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.712000 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/45ebd02f-dc19-48a5-8309-bdbf11a8a7a4-etcd-service-ca\") pod \"etcd-operator-b45778765-7sxxj\" (UID: \"45ebd02f-dc19-48a5-8309-bdbf11a8a7a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7sxxj" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.708855 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e20a512-8f34-4d21-be45-e749f7a6cc87-trusted-ca-bundle\") pod \"console-f9d7485db-csl6r\" (UID: \"5e20a512-8f34-4d21-be45-e749f7a6cc87\") " pod="openshift-console/console-f9d7485db-csl6r" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.712147 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.712379 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/14860b75-6fa1-43b5-ba20-d731531dba3b-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-ws79r\" (UID: \"14860b75-6fa1-43b5-ba20-d731531dba3b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ws79r" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.712521 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f4cc87b-f2fd-4c61-82fb-f362b81445d7-config\") pod \"openshift-apiserver-operator-796bbdcf4f-bjc5w\" (UID: \"8f4cc87b-f2fd-4c61-82fb-f362b81445d7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bjc5w" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.712708 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5e20a512-8f34-4d21-be45-e749f7a6cc87-console-config\") pod \"console-f9d7485db-csl6r\" (UID: \"5e20a512-8f34-4d21-be45-e749f7a6cc87\") " pod="openshift-console/console-f9d7485db-csl6r" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.712836 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7b200d8c-28e7-494b-99ec-893f1d9ba3ca-auth-proxy-config\") pod \"machine-config-operator-74547568cd-cp2n5\" (UID: \"7b200d8c-28e7-494b-99ec-893f1d9ba3ca\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cp2n5" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.713442 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/d89bf2d0-8109-40be-b1ce-26ee92a58d81-available-featuregates\") pod \"openshift-config-operator-7777fb866f-ws6r9\" (UID: \"d89bf2d0-8109-40be-b1ce-26ee92a58d81\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ws6r9" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.713995 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0dae89fc-445f-4531-8928-29b334636522-audit-policies\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.714153 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3bc0bb1d-5b22-4e74-8aa8-6554b77a2b73-service-ca-bundle\") pod \"authentication-operator-69f744f599-qn4g8\" (UID: \"3bc0bb1d-5b22-4e74-8aa8-6554b77a2b73\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qn4g8" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.714277 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3bc0bb1d-5b22-4e74-8aa8-6554b77a2b73-config\") pod \"authentication-operator-69f744f599-qn4g8\" (UID: \"3bc0bb1d-5b22-4e74-8aa8-6554b77a2b73\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qn4g8" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.708869 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0dae89fc-445f-4531-8928-29b334636522-audit-dir\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.710262 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/943c9c2f-24f5-408c-8e8d-c344253fc8f6-config\") pod \"machine-approver-56656f9798-vwp4c\" (UID: \"943c9c2f-24f5-408c-8e8d-c344253fc8f6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwp4c" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.714575 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a4ef12d4-a717-4afa-b3c7-179ea0f7f16e-trusted-ca\") pod \"console-operator-58897d9998-rt5vd\" (UID: \"a4ef12d4-a717-4afa-b3c7-179ea0f7f16e\") " pod="openshift-console-operator/console-operator-58897d9998-rt5vd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.714612 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/49347426-bca5-4fde-8b87-21dc8fc486d8-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-kmpsc\" (UID: \"49347426-bca5-4fde-8b87-21dc8fc486d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kmpsc" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.710629 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.714990 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5449c948-487e-426d-b151-31e8079576f3-client-ca\") pod \"route-controller-manager-6576b87f9c-wtn79\" (UID: \"5449c948-487e-426d-b151-31e8079576f3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wtn79" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.715507 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/943c9c2f-24f5-408c-8e8d-c344253fc8f6-auth-proxy-config\") pod \"machine-approver-56656f9798-vwp4c\" (UID: \"943c9c2f-24f5-408c-8e8d-c344253fc8f6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwp4c" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.715576 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/038209e7-3fdb-490b-976e-f1791bff29d8-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-cwz9p\" (UID: \"038209e7-3fdb-490b-976e-f1791bff29d8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cwz9p" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.715816 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5e20a512-8f34-4d21-be45-e749f7a6cc87-service-ca\") pod \"console-f9d7485db-csl6r\" (UID: \"5e20a512-8f34-4d21-be45-e749f7a6cc87\") " pod="openshift-console/console-f9d7485db-csl6r" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.715836 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/3bc0bb1d-5b22-4e74-8aa8-6554b77a2b73-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-qn4g8\" (UID: \"3bc0bb1d-5b22-4e74-8aa8-6554b77a2b73\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qn4g8" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.715856 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.716499 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.716890 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/45ebd02f-dc19-48a5-8309-bdbf11a8a7a4-etcd-client\") pod \"etcd-operator-b45778765-7sxxj\" (UID: \"45ebd02f-dc19-48a5-8309-bdbf11a8a7a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7sxxj" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.717172 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/49347426-bca5-4fde-8b87-21dc8fc486d8-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-kmpsc\" (UID: \"49347426-bca5-4fde-8b87-21dc8fc486d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kmpsc" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.717727 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5e20a512-8f34-4d21-be45-e749f7a6cc87-console-oauth-config\") pod \"console-f9d7485db-csl6r\" (UID: \"5e20a512-8f34-4d21-be45-e749f7a6cc87\") " pod="openshift-console/console-f9d7485db-csl6r" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.717741 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5e20a512-8f34-4d21-be45-e749f7a6cc87-oauth-serving-cert\") pod \"console-f9d7485db-csl6r\" (UID: \"5e20a512-8f34-4d21-be45-e749f7a6cc87\") " pod="openshift-console/console-f9d7485db-csl6r" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.718448 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.718674 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5449c948-487e-426d-b151-31e8079576f3-serving-cert\") pod \"route-controller-manager-6576b87f9c-wtn79\" (UID: \"5449c948-487e-426d-b151-31e8079576f3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wtn79" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.718708 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3bc0bb1d-5b22-4e74-8aa8-6554b77a2b73-serving-cert\") pod \"authentication-operator-69f744f599-qn4g8\" (UID: \"3bc0bb1d-5b22-4e74-8aa8-6554b77a2b73\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qn4g8" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.718776 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/943c9c2f-24f5-408c-8e8d-c344253fc8f6-machine-approver-tls\") pod \"machine-approver-56656f9798-vwp4c\" (UID: \"943c9c2f-24f5-408c-8e8d-c344253fc8f6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwp4c" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.719304 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.719933 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.720027 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.720295 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f4cc87b-f2fd-4c61-82fb-f362b81445d7-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-bjc5w\" (UID: \"8f4cc87b-f2fd-4c61-82fb-f362b81445d7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bjc5w" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.720359 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/b5fca2ad-c70a-43e2-aff1-f180617a94f8-metrics-tls\") pod \"dns-operator-744455d44c-9jjwn\" (UID: \"b5fca2ad-c70a-43e2-aff1-f180617a94f8\") " pod="openshift-dns-operator/dns-operator-744455d44c-9jjwn" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.720711 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5e20a512-8f34-4d21-be45-e749f7a6cc87-console-serving-cert\") pod \"console-f9d7485db-csl6r\" (UID: \"5e20a512-8f34-4d21-be45-e749f7a6cc87\") " pod="openshift-console/console-f9d7485db-csl6r" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.720900 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.723204 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/45ebd02f-dc19-48a5-8309-bdbf11a8a7a4-serving-cert\") pod \"etcd-operator-b45778765-7sxxj\" (UID: \"45ebd02f-dc19-48a5-8309-bdbf11a8a7a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7sxxj" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.723796 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a4ef12d4-a717-4afa-b3c7-179ea0f7f16e-serving-cert\") pod \"console-operator-58897d9998-rt5vd\" (UID: \"a4ef12d4-a717-4afa-b3c7-179ea0f7f16e\") " pod="openshift-console-operator/console-operator-58897d9998-rt5vd" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.725044 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.749665 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.759398 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b173cdfc-6a31-4d2e-93cd-a04eb1c68877-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rw8ng\" (UID: \"b173cdfc-6a31-4d2e-93cd-a04eb1c68877\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rw8ng" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.764200 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.771501 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b173cdfc-6a31-4d2e-93cd-a04eb1c68877-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rw8ng\" (UID: \"b173cdfc-6a31-4d2e-93cd-a04eb1c68877\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rw8ng" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.784326 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.805852 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.812352 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0761866e-d6ab-40c9-a776-8c7fb2683fad-service-ca-bundle\") pod \"router-default-5444994796-2fvpq\" (UID: \"0761866e-d6ab-40c9-a776-8c7fb2683fad\") " pod="openshift-ingress/router-default-5444994796-2fvpq" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.812390 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwxhr\" (UniqueName: \"kubernetes.io/projected/0761866e-d6ab-40c9-a776-8c7fb2683fad-kube-api-access-pwxhr\") pod \"router-default-5444994796-2fvpq\" (UID: \"0761866e-d6ab-40c9-a776-8c7fb2683fad\") " pod="openshift-ingress/router-default-5444994796-2fvpq" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.812430 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c506a84b-4ea6-477e-88de-fceec3a95f31-config-volume\") pod \"collect-profiles-29335905-kdpld\" (UID: \"c506a84b-4ea6-477e-88de-fceec3a95f31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-kdpld" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.812465 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0761866e-d6ab-40c9-a776-8c7fb2683fad-metrics-certs\") pod \"router-default-5444994796-2fvpq\" (UID: \"0761866e-d6ab-40c9-a776-8c7fb2683fad\") " pod="openshift-ingress/router-default-5444994796-2fvpq" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.812496 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lzw95\" (UniqueName: \"kubernetes.io/projected/23902dde-0c8d-4d54-99f4-0ee4e19a1a50-kube-api-access-lzw95\") pod \"machine-config-server-fsc8x\" (UID: \"23902dde-0c8d-4d54-99f4-0ee4e19a1a50\") " pod="openshift-machine-config-operator/machine-config-server-fsc8x" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.812511 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/47a99167-adc4-40b4-9925-778c670ad55f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-g4q8m\" (UID: \"47a99167-adc4-40b4-9925-778c670ad55f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-g4q8m" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.812555 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/23902dde-0c8d-4d54-99f4-0ee4e19a1a50-certs\") pod \"machine-config-server-fsc8x\" (UID: \"23902dde-0c8d-4d54-99f4-0ee4e19a1a50\") " pod="openshift-machine-config-operator/machine-config-server-fsc8x" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.812611 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/0761866e-d6ab-40c9-a776-8c7fb2683fad-default-certificate\") pod \"router-default-5444994796-2fvpq\" (UID: \"0761866e-d6ab-40c9-a776-8c7fb2683fad\") " pod="openshift-ingress/router-default-5444994796-2fvpq" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.812635 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fwwhz\" (UniqueName: \"kubernetes.io/projected/c506a84b-4ea6-477e-88de-fceec3a95f31-kube-api-access-fwwhz\") pod \"collect-profiles-29335905-kdpld\" (UID: \"c506a84b-4ea6-477e-88de-fceec3a95f31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-kdpld" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.812650 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c506a84b-4ea6-477e-88de-fceec3a95f31-secret-volume\") pod \"collect-profiles-29335905-kdpld\" (UID: \"c506a84b-4ea6-477e-88de-fceec3a95f31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-kdpld" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.812696 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/0761866e-d6ab-40c9-a776-8c7fb2683fad-stats-auth\") pod \"router-default-5444994796-2fvpq\" (UID: \"0761866e-d6ab-40c9-a776-8c7fb2683fad\") " pod="openshift-ingress/router-default-5444994796-2fvpq" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.812723 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/23902dde-0c8d-4d54-99f4-0ee4e19a1a50-node-bootstrap-token\") pod \"machine-config-server-fsc8x\" (UID: \"23902dde-0c8d-4d54-99f4-0ee4e19a1a50\") " pod="openshift-machine-config-operator/machine-config-server-fsc8x" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.812826 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxdkj\" (UniqueName: \"kubernetes.io/projected/47a99167-adc4-40b4-9925-778c670ad55f-kube-api-access-gxdkj\") pod \"multus-admission-controller-857f4d67dd-g4q8m\" (UID: \"47a99167-adc4-40b4-9925-778c670ad55f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-g4q8m" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.824585 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.832124 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/ac8fc0d4-aa55-4974-a506-7ccd2713f31a-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-gwrxz\" (UID: \"ac8fc0d4-aa55-4974-a506-7ccd2713f31a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gwrxz" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.844203 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.846403 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ac8fc0d4-aa55-4974-a506-7ccd2713f31a-config\") pod \"kube-apiserver-operator-766d6c64bb-gwrxz\" (UID: \"ac8fc0d4-aa55-4974-a506-7ccd2713f31a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gwrxz" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.865352 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.868522 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/038209e7-3fdb-490b-976e-f1791bff29d8-proxy-tls\") pod \"machine-config-controller-84d6567774-cwz9p\" (UID: \"038209e7-3fdb-490b-976e-f1791bff29d8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cwz9p" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.884275 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.903579 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.924028 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.933546 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/7fa0bf77-28c4-492d-a4d4-b0a0ec401cf0-metrics-tls\") pod \"ingress-operator-5b745b69d9-n6qrf\" (UID: \"7fa0bf77-28c4-492d-a4d4-b0a0ec401cf0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n6qrf" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.943614 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.964162 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.992803 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Oct 11 03:53:31 crc kubenswrapper[4967]: I1011 03:53:31.995961 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/7fa0bf77-28c4-492d-a4d4-b0a0ec401cf0-trusted-ca\") pod \"ingress-operator-5b745b69d9-n6qrf\" (UID: \"7fa0bf77-28c4-492d-a4d4-b0a0ec401cf0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n6qrf" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.004014 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.025157 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.031154 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7b200d8c-28e7-494b-99ec-893f1d9ba3ca-proxy-tls\") pod \"machine-config-operator-74547568cd-cp2n5\" (UID: \"7b200d8c-28e7-494b-99ec-893f1d9ba3ca\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cp2n5" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.044519 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.065908 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.074891 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/7b200d8c-28e7-494b-99ec-893f1d9ba3ca-images\") pod \"machine-config-operator-74547568cd-cp2n5\" (UID: \"7b200d8c-28e7-494b-99ec-893f1d9ba3ca\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cp2n5" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.084949 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.104762 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.111318 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/c3a20ddb-cfaa-46ac-8aed-4db9de3e32b6-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-rv6n2\" (UID: \"c3a20ddb-cfaa-46ac-8aed-4db9de3e32b6\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rv6n2" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.125843 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.146871 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.165294 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.176728 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/89f0e615-1fde-483d-9b38-83287f104360-profile-collector-cert\") pod \"olm-operator-6b444d44fb-bxjls\" (UID: \"89f0e615-1fde-483d-9b38-83287f104360\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bxjls" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.178238 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c506a84b-4ea6-477e-88de-fceec3a95f31-secret-volume\") pod \"collect-profiles-29335905-kdpld\" (UID: \"c506a84b-4ea6-477e-88de-fceec3a95f31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-kdpld" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.184840 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.203961 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.211966 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/89f0e615-1fde-483d-9b38-83287f104360-srv-cert\") pod \"olm-operator-6b444d44fb-bxjls\" (UID: \"89f0e615-1fde-483d-9b38-83287f104360\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bxjls" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.224657 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.244115 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.265182 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.285182 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.305330 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.324425 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.345898 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.364777 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.385157 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.405001 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.424823 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.444949 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.466016 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.477706 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/092da80d-ee9c-4e88-b3e2-58bb3dd77f9b-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-tr7z7\" (UID: \"092da80d-ee9c-4e88-b3e2-58bb3dd77f9b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tr7z7" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.484651 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.505141 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.514364 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/092da80d-ee9c-4e88-b3e2-58bb3dd77f9b-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-tr7z7\" (UID: \"092da80d-ee9c-4e88-b3e2-58bb3dd77f9b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tr7z7" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.526157 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.544937 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.562994 4967 request.go:700] Waited for 1.016649323s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-ingress/secrets?fieldSelector=metadata.name%3Drouter-dockercfg-zdk86&limit=500&resourceVersion=0 Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.564674 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.585214 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.598103 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/0761866e-d6ab-40c9-a776-8c7fb2683fad-default-certificate\") pod \"router-default-5444994796-2fvpq\" (UID: \"0761866e-d6ab-40c9-a776-8c7fb2683fad\") " pod="openshift-ingress/router-default-5444994796-2fvpq" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.605625 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.619618 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/0761866e-d6ab-40c9-a776-8c7fb2683fad-stats-auth\") pod \"router-default-5444994796-2fvpq\" (UID: \"0761866e-d6ab-40c9-a776-8c7fb2683fad\") " pod="openshift-ingress/router-default-5444994796-2fvpq" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.627612 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.637219 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/0761866e-d6ab-40c9-a776-8c7fb2683fad-metrics-certs\") pod \"router-default-5444994796-2fvpq\" (UID: \"0761866e-d6ab-40c9-a776-8c7fb2683fad\") " pod="openshift-ingress/router-default-5444994796-2fvpq" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.644905 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.653527 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0761866e-d6ab-40c9-a776-8c7fb2683fad-service-ca-bundle\") pod \"router-default-5444994796-2fvpq\" (UID: \"0761866e-d6ab-40c9-a776-8c7fb2683fad\") " pod="openshift-ingress/router-default-5444994796-2fvpq" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.664712 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.685606 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.694688 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/3883197f-52dd-465a-82bb-c4dd32b3d090-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-rmkqj\" (UID: \"3883197f-52dd-465a-82bb-c4dd32b3d090\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rmkqj" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.704412 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.710611 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3883197f-52dd-465a-82bb-c4dd32b3d090-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-rmkqj\" (UID: \"3883197f-52dd-465a-82bb-c4dd32b3d090\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rmkqj" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.725180 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.745445 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.765591 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.785182 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.796440 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/47a99167-adc4-40b4-9925-778c670ad55f-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-g4q8m\" (UID: \"47a99167-adc4-40b4-9925-778c670ad55f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-g4q8m" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.805291 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Oct 11 03:53:32 crc kubenswrapper[4967]: E1011 03:53:32.813152 4967 secret.go:188] Couldn't get secret openshift-machine-config-operator/node-bootstrapper-token: failed to sync secret cache: timed out waiting for the condition Oct 11 03:53:32 crc kubenswrapper[4967]: E1011 03:53:32.813199 4967 configmap.go:193] Couldn't get configMap openshift-operator-lifecycle-manager/collect-profiles-config: failed to sync configmap cache: timed out waiting for the condition Oct 11 03:53:32 crc kubenswrapper[4967]: E1011 03:53:32.813161 4967 secret.go:188] Couldn't get secret openshift-machine-config-operator/machine-config-server-tls: failed to sync secret cache: timed out waiting for the condition Oct 11 03:53:32 crc kubenswrapper[4967]: E1011 03:53:32.813247 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/23902dde-0c8d-4d54-99f4-0ee4e19a1a50-node-bootstrap-token podName:23902dde-0c8d-4d54-99f4-0ee4e19a1a50 nodeName:}" failed. No retries permitted until 2025-10-11 03:53:33.313224973 +0000 UTC m=+141.276434006 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "node-bootstrap-token" (UniqueName: "kubernetes.io/secret/23902dde-0c8d-4d54-99f4-0ee4e19a1a50-node-bootstrap-token") pod "machine-config-server-fsc8x" (UID: "23902dde-0c8d-4d54-99f4-0ee4e19a1a50") : failed to sync secret cache: timed out waiting for the condition Oct 11 03:53:32 crc kubenswrapper[4967]: E1011 03:53:32.813429 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/c506a84b-4ea6-477e-88de-fceec3a95f31-config-volume podName:c506a84b-4ea6-477e-88de-fceec3a95f31 nodeName:}" failed. No retries permitted until 2025-10-11 03:53:33.313372547 +0000 UTC m=+141.276581510 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config-volume" (UniqueName: "kubernetes.io/configmap/c506a84b-4ea6-477e-88de-fceec3a95f31-config-volume") pod "collect-profiles-29335905-kdpld" (UID: "c506a84b-4ea6-477e-88de-fceec3a95f31") : failed to sync configmap cache: timed out waiting for the condition Oct 11 03:53:32 crc kubenswrapper[4967]: E1011 03:53:32.813464 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/23902dde-0c8d-4d54-99f4-0ee4e19a1a50-certs podName:23902dde-0c8d-4d54-99f4-0ee4e19a1a50 nodeName:}" failed. No retries permitted until 2025-10-11 03:53:33.313448569 +0000 UTC m=+141.276657542 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "certs" (UniqueName: "kubernetes.io/secret/23902dde-0c8d-4d54-99f4-0ee4e19a1a50-certs") pod "machine-config-server-fsc8x" (UID: "23902dde-0c8d-4d54-99f4-0ee4e19a1a50") : failed to sync secret cache: timed out waiting for the condition Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.844814 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.865519 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.885000 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.905611 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.934496 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.944376 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.964945 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Oct 11 03:53:32 crc kubenswrapper[4967]: I1011 03:53:32.984929 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.004459 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.024825 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.045352 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.065040 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.084426 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.105464 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.125442 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.145126 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.164986 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.185240 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.225304 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9w2rs\" (UniqueName: \"kubernetes.io/projected/9b3efa80-8370-4353-b688-4701c4deec8f-kube-api-access-9w2rs\") pod \"machine-api-operator-5694c8668f-mrsjk\" (UID: \"9b3efa80-8370-4353-b688-4701c4deec8f\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-mrsjk" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.258455 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qg482\" (UniqueName: \"kubernetes.io/projected/454804e7-0d4d-4536-8544-87a5b6368f70-kube-api-access-qg482\") pod \"apiserver-7bbb656c7d-nmzrx\" (UID: \"454804e7-0d4d-4536-8544-87a5b6368f70\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.264882 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.269788 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-mrsjk" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.270997 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8vnp\" (UniqueName: \"kubernetes.io/projected/1c77fe9a-d2e1-465b-8600-9f7f2d1fa206-kube-api-access-j8vnp\") pod \"controller-manager-879f6c89f-92sz4\" (UID: \"1c77fe9a-d2e1-465b-8600-9f7f2d1fa206\") " pod="openshift-controller-manager/controller-manager-879f6c89f-92sz4" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.285154 4967 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.290344 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.305347 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.341676 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c506a84b-4ea6-477e-88de-fceec3a95f31-config-volume\") pod \"collect-profiles-29335905-kdpld\" (UID: \"c506a84b-4ea6-477e-88de-fceec3a95f31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-kdpld" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.341872 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/23902dde-0c8d-4d54-99f4-0ee4e19a1a50-certs\") pod \"machine-config-server-fsc8x\" (UID: \"23902dde-0c8d-4d54-99f4-0ee4e19a1a50\") " pod="openshift-machine-config-operator/machine-config-server-fsc8x" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.342110 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/23902dde-0c8d-4d54-99f4-0ee4e19a1a50-node-bootstrap-token\") pod \"machine-config-server-fsc8x\" (UID: \"23902dde-0c8d-4d54-99f4-0ee4e19a1a50\") " pod="openshift-machine-config-operator/machine-config-server-fsc8x" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.344005 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c506a84b-4ea6-477e-88de-fceec3a95f31-config-volume\") pod \"collect-profiles-29335905-kdpld\" (UID: \"c506a84b-4ea6-477e-88de-fceec3a95f31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-kdpld" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.345381 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.353327 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hk288\" (UniqueName: \"kubernetes.io/projected/efe9c94d-8f97-4274-a494-a5ecf5b88cec-kube-api-access-hk288\") pod \"apiserver-76f77b778f-nxdtr\" (UID: \"efe9c94d-8f97-4274-a494-a5ecf5b88cec\") " pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.366006 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.378586 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/23902dde-0c8d-4d54-99f4-0ee4e19a1a50-certs\") pod \"machine-config-server-fsc8x\" (UID: \"23902dde-0c8d-4d54-99f4-0ee4e19a1a50\") " pod="openshift-machine-config-operator/machine-config-server-fsc8x" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.384689 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.398249 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/23902dde-0c8d-4d54-99f4-0ee4e19a1a50-node-bootstrap-token\") pod \"machine-config-server-fsc8x\" (UID: \"23902dde-0c8d-4d54-99f4-0ee4e19a1a50\") " pod="openshift-machine-config-operator/machine-config-server-fsc8x" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.406480 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.425377 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.446743 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.484847 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.504916 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.525489 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.529550 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-92sz4" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.549379 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.563670 4967 request.go:700] Waited for 1.85724217s due to client-side throttling, not priority and fairness, request: POST:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication-operator/serviceaccounts/authentication-operator/token Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.574883 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.576337 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-mrsjk"] Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.577847 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx"] Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.580868 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rwhh8\" (UniqueName: \"kubernetes.io/projected/3bc0bb1d-5b22-4e74-8aa8-6554b77a2b73-kube-api-access-rwhh8\") pod \"authentication-operator-69f744f599-qn4g8\" (UID: \"3bc0bb1d-5b22-4e74-8aa8-6554b77a2b73\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-qn4g8" Oct 11 03:53:33 crc kubenswrapper[4967]: W1011 03:53:33.589759 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod454804e7_0d4d_4536_8544_87a5b6368f70.slice/crio-a3e660d8a95d16aea11ecefb9f58831c58e187a75b3c57a6ec27e614f3dfc49d WatchSource:0}: Error finding container a3e660d8a95d16aea11ecefb9f58831c58e187a75b3c57a6ec27e614f3dfc49d: Status 404 returned error can't find the container with id a3e660d8a95d16aea11ecefb9f58831c58e187a75b3c57a6ec27e614f3dfc49d Oct 11 03:53:33 crc kubenswrapper[4967]: W1011 03:53:33.591397 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod9b3efa80_8370_4353_b688_4701c4deec8f.slice/crio-ac0ddf6d1149c137adcf03a36d895900d3dbb86b6c7d825bb209d1172d681456 WatchSource:0}: Error finding container ac0ddf6d1149c137adcf03a36d895900d3dbb86b6c7d825bb209d1172d681456: Status 404 returned error can't find the container with id ac0ddf6d1149c137adcf03a36d895900d3dbb86b6c7d825bb209d1172d681456 Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.600682 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xlsbr\" (UniqueName: \"kubernetes.io/projected/a4ef12d4-a717-4afa-b3c7-179ea0f7f16e-kube-api-access-xlsbr\") pod \"console-operator-58897d9998-rt5vd\" (UID: \"a4ef12d4-a717-4afa-b3c7-179ea0f7f16e\") " pod="openshift-console-operator/console-operator-58897d9998-rt5vd" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.662158 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pgqsl\" (UniqueName: \"kubernetes.io/projected/b5fca2ad-c70a-43e2-aff1-f180617a94f8-kube-api-access-pgqsl\") pod \"dns-operator-744455d44c-9jjwn\" (UID: \"b5fca2ad-c70a-43e2-aff1-f180617a94f8\") " pod="openshift-dns-operator/dns-operator-744455d44c-9jjwn" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.662555 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2gb75\" (UniqueName: \"kubernetes.io/projected/038209e7-3fdb-490b-976e-f1791bff29d8-kube-api-access-2gb75\") pod \"machine-config-controller-84d6567774-cwz9p\" (UID: \"038209e7-3fdb-490b-976e-f1791bff29d8\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cwz9p" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.665886 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t6s92\" (UniqueName: \"kubernetes.io/projected/3883197f-52dd-465a-82bb-c4dd32b3d090-kube-api-access-t6s92\") pod \"kube-storage-version-migrator-operator-b67b599dd-rmkqj\" (UID: \"3883197f-52dd-465a-82bb-c4dd32b3d090\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rmkqj" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.705900 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tdtmh\" (UniqueName: \"kubernetes.io/projected/d6b9a450-fd86-47a3-9c0e-f61aa7a24377-kube-api-access-tdtmh\") pod \"downloads-7954f5f757-dtj88\" (UID: \"d6b9a450-fd86-47a3-9c0e-f61aa7a24377\") " pod="openshift-console/downloads-7954f5f757-dtj88" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.711537 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zwlq2\" (UniqueName: \"kubernetes.io/projected/5e20a512-8f34-4d21-be45-e749f7a6cc87-kube-api-access-zwlq2\") pod \"console-f9d7485db-csl6r\" (UID: \"5e20a512-8f34-4d21-be45-e749f7a6cc87\") " pod="openshift-console/console-f9d7485db-csl6r" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.728832 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wpmjz\" (UniqueName: \"kubernetes.io/projected/943c9c2f-24f5-408c-8e8d-c344253fc8f6-kube-api-access-wpmjz\") pod \"machine-approver-56656f9798-vwp4c\" (UID: \"943c9c2f-24f5-408c-8e8d-c344253fc8f6\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwp4c" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.740732 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xtvtk\" (UniqueName: \"kubernetes.io/projected/0dae89fc-445f-4531-8928-29b334636522-kube-api-access-xtvtk\") pod \"oauth-openshift-558db77b4-6sjfd\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.747533 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-rt5vd" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.758405 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/7fa0bf77-28c4-492d-a4d4-b0a0ec401cf0-bound-sa-token\") pod \"ingress-operator-5b745b69d9-n6qrf\" (UID: \"7fa0bf77-28c4-492d-a4d4-b0a0ec401cf0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n6qrf" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.767330 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-qn4g8" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.772240 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-92sz4"] Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.778978 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/ac8fc0d4-aa55-4974-a506-7ccd2713f31a-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-gwrxz\" (UID: \"ac8fc0d4-aa55-4974-a506-7ccd2713f31a\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gwrxz" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.783525 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-9jjwn" Oct 11 03:53:33 crc kubenswrapper[4967]: W1011 03:53:33.791212 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1c77fe9a_d2e1_465b_8600_9f7f2d1fa206.slice/crio-8675633416cc7576fe45dc4505545a7520fc1778845f60d19e362df7a3e1480a WatchSource:0}: Error finding container 8675633416cc7576fe45dc4505545a7520fc1778845f60d19e362df7a3e1480a: Status 404 returned error can't find the container with id 8675633416cc7576fe45dc4505545a7520fc1778845f60d19e362df7a3e1480a Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.800109 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.801565 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xs6pj\" (UniqueName: \"kubernetes.io/projected/45ebd02f-dc19-48a5-8309-bdbf11a8a7a4-kube-api-access-xs6pj\") pod \"etcd-operator-b45778765-7sxxj\" (UID: \"45ebd02f-dc19-48a5-8309-bdbf11a8a7a4\") " pod="openshift-etcd-operator/etcd-operator-b45778765-7sxxj" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.809323 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-csl6r" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.824421 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-nxdtr"] Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.829011 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wk44b\" (UniqueName: \"kubernetes.io/projected/49347426-bca5-4fde-8b87-21dc8fc486d8-kube-api-access-wk44b\") pod \"cluster-image-registry-operator-dc59b4c8b-kmpsc\" (UID: \"49347426-bca5-4fde-8b87-21dc8fc486d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kmpsc" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.845728 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bmclp\" (UniqueName: \"kubernetes.io/projected/8f4cc87b-f2fd-4c61-82fb-f362b81445d7-kube-api-access-bmclp\") pod \"openshift-apiserver-operator-796bbdcf4f-bjc5w\" (UID: \"8f4cc87b-f2fd-4c61-82fb-f362b81445d7\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bjc5w" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.860317 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gwrxz" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.871987 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cwz9p" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.874721 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/49347426-bca5-4fde-8b87-21dc8fc486d8-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-kmpsc\" (UID: \"49347426-bca5-4fde-8b87-21dc8fc486d8\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kmpsc" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.880560 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qwv5t\" (UniqueName: \"kubernetes.io/projected/89f0e615-1fde-483d-9b38-83287f104360-kube-api-access-qwv5t\") pod \"olm-operator-6b444d44fb-bxjls\" (UID: \"89f0e615-1fde-483d-9b38-83287f104360\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bxjls" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.895140 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rmkqj" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.901503 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6j54n\" (UniqueName: \"kubernetes.io/projected/7b200d8c-28e7-494b-99ec-893f1d9ba3ca-kube-api-access-6j54n\") pod \"machine-config-operator-74547568cd-cp2n5\" (UID: \"7b200d8c-28e7-494b-99ec-893f1d9ba3ca\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cp2n5" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.909565 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bxjls" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.919054 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-shzqz\" (UniqueName: \"kubernetes.io/projected/7fa0bf77-28c4-492d-a4d4-b0a0ec401cf0-kube-api-access-shzqz\") pod \"ingress-operator-5b745b69d9-n6qrf\" (UID: \"7fa0bf77-28c4-492d-a4d4-b0a0ec401cf0\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n6qrf" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.940302 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-rt5vd"] Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.944415 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kp59m\" (UniqueName: \"kubernetes.io/projected/092da80d-ee9c-4e88-b3e2-58bb3dd77f9b-kube-api-access-kp59m\") pod \"openshift-controller-manager-operator-756b6f6bc6-tr7z7\" (UID: \"092da80d-ee9c-4e88-b3e2-58bb3dd77f9b\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tr7z7" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.948343 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tr7z7" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.951520 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwp4c" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.961575 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v7dwc\" (UniqueName: \"kubernetes.io/projected/c3a20ddb-cfaa-46ac-8aed-4db9de3e32b6-kube-api-access-v7dwc\") pod \"control-plane-machine-set-operator-78cbb6b69f-rv6n2\" (UID: \"c3a20ddb-cfaa-46ac-8aed-4db9de3e32b6\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rv6n2" Oct 11 03:53:33 crc kubenswrapper[4967]: W1011 03:53:33.971325 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda4ef12d4_a717_4afa_b3c7_179ea0f7f16e.slice/crio-72a90e18a41753bb9c46909491a93c5f6392facd793acf6b50939acc7d2d48c8 WatchSource:0}: Error finding container 72a90e18a41753bb9c46909491a93c5f6392facd793acf6b50939acc7d2d48c8: Status 404 returned error can't find the container with id 72a90e18a41753bb9c46909491a93c5f6392facd793acf6b50939acc7d2d48c8 Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.983334 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-dtj88" Oct 11 03:53:33 crc kubenswrapper[4967]: I1011 03:53:33.990888 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-9jjwn"] Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:33.999973 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6vbsf\" (UniqueName: \"kubernetes.io/projected/81c6feb2-e6ea-43f0-834d-499bea65b4aa-kube-api-access-6vbsf\") pod \"migrator-59844c95c7-gbqfq\" (UID: \"81c6feb2-e6ea-43f0-834d-499bea65b4aa\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gbqfq" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.000271 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wf8bd\" (UniqueName: \"kubernetes.io/projected/d89bf2d0-8109-40be-b1ce-26ee92a58d81-kube-api-access-wf8bd\") pod \"openshift-config-operator-7777fb866f-ws6r9\" (UID: \"d89bf2d0-8109-40be-b1ce-26ee92a58d81\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-ws6r9" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.012773 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-qn4g8"] Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.026825 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sspzl\" (UniqueName: \"kubernetes.io/projected/14860b75-6fa1-43b5-ba20-d731531dba3b-kube-api-access-sspzl\") pod \"cluster-samples-operator-665b6dd947-ws79r\" (UID: \"14860b75-6fa1-43b5-ba20-d731531dba3b\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ws79r" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.042730 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qvzlg\" (UniqueName: \"kubernetes.io/projected/5449c948-487e-426d-b151-31e8079576f3-kube-api-access-qvzlg\") pod \"route-controller-manager-6576b87f9c-wtn79\" (UID: \"5449c948-487e-426d-b151-31e8079576f3\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wtn79" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.046793 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bjc5w" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.056258 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kmpsc" Oct 11 03:53:34 crc kubenswrapper[4967]: W1011 03:53:34.060116 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5fca2ad_c70a_43e2_aff1_f180617a94f8.slice/crio-f8a91aa92c40c5527fb2f37654d1107888983f199b16a060acde35cb3e736d2b WatchSource:0}: Error finding container f8a91aa92c40c5527fb2f37654d1107888983f199b16a060acde35cb3e736d2b: Status 404 returned error can't find the container with id f8a91aa92c40c5527fb2f37654d1107888983f199b16a060acde35cb3e736d2b Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.067822 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/b173cdfc-6a31-4d2e-93cd-a04eb1c68877-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-rw8ng\" (UID: \"b173cdfc-6a31-4d2e-93cd-a04eb1c68877\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rw8ng" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.075861 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ws6r9" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.078423 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwxhr\" (UniqueName: \"kubernetes.io/projected/0761866e-d6ab-40c9-a776-8c7fb2683fad-kube-api-access-pwxhr\") pod \"router-default-5444994796-2fvpq\" (UID: \"0761866e-d6ab-40c9-a776-8c7fb2683fad\") " pod="openshift-ingress/router-default-5444994796-2fvpq" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.091704 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-7sxxj" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.100763 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lzw95\" (UniqueName: \"kubernetes.io/projected/23902dde-0c8d-4d54-99f4-0ee4e19a1a50-kube-api-access-lzw95\") pod \"machine-config-server-fsc8x\" (UID: \"23902dde-0c8d-4d54-99f4-0ee4e19a1a50\") " pod="openshift-machine-config-operator/machine-config-server-fsc8x" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.104013 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6sjfd"] Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.120752 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fwwhz\" (UniqueName: \"kubernetes.io/projected/c506a84b-4ea6-477e-88de-fceec3a95f31-kube-api-access-fwwhz\") pod \"collect-profiles-29335905-kdpld\" (UID: \"c506a84b-4ea6-477e-88de-fceec3a95f31\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-kdpld" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.150751 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxdkj\" (UniqueName: \"kubernetes.io/projected/47a99167-adc4-40b4-9925-778c670ad55f-kube-api-access-gxdkj\") pod \"multus-admission-controller-857f4d67dd-g4q8m\" (UID: \"47a99167-adc4-40b4-9925-778c670ad55f\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-g4q8m" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.153793 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rw8ng" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.178996 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n6qrf" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.190412 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cp2n5" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.203010 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rv6n2" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.218928 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gbqfq" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.226910 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wtn79" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.256456 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-2fvpq" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.257915 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n5dnf\" (UniqueName: \"kubernetes.io/projected/e6b311ab-c933-4950-9104-8071d75d77ac-kube-api-access-n5dnf\") pod \"packageserver-d55dfcdfc-wnjpj\" (UID: \"e6b311ab-c933-4950-9104-8071d75d77ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnjpj" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.257952 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sb5vr\" (UniqueName: \"kubernetes.io/projected/3c23d015-ef00-4a30-b2da-dc15d140a773-kube-api-access-sb5vr\") pod \"dns-default-d8x8n\" (UID: \"3c23d015-ef00-4a30-b2da-dc15d140a773\") " pod="openshift-dns/dns-default-d8x8n" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.257994 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w7w9f\" (UniqueName: \"kubernetes.io/projected/de60a4e8-2c49-4b5b-a28d-b782c0852911-kube-api-access-w7w9f\") pod \"service-ca-9c57cc56f-4xkx2\" (UID: \"de60a4e8-2c49-4b5b-a28d-b782c0852911\") " pod="openshift-service-ca/service-ca-9c57cc56f-4xkx2" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.258120 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e6b311ab-c933-4950-9104-8071d75d77ac-apiservice-cert\") pod \"packageserver-d55dfcdfc-wnjpj\" (UID: \"e6b311ab-c933-4950-9104-8071d75d77ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnjpj" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.258163 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/de60a4e8-2c49-4b5b-a28d-b782c0852911-signing-cabundle\") pod \"service-ca-9c57cc56f-4xkx2\" (UID: \"de60a4e8-2c49-4b5b-a28d-b782c0852911\") " pod="openshift-service-ca/service-ca-9c57cc56f-4xkx2" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.258181 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/e1a3f43f-63fe-4816-8414-2d7e4c63be57-socket-dir\") pod \"csi-hostpathplugin-dnnwg\" (UID: \"e1a3f43f-63fe-4816-8414-2d7e4c63be57\") " pod="hostpath-provisioner/csi-hostpathplugin-dnnwg" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.258205 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e6b311ab-c933-4950-9104-8071d75d77ac-webhook-cert\") pod \"packageserver-d55dfcdfc-wnjpj\" (UID: \"e6b311ab-c933-4950-9104-8071d75d77ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnjpj" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.258254 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/e6b311ab-c933-4950-9104-8071d75d77ac-tmpfs\") pod \"packageserver-d55dfcdfc-wnjpj\" (UID: \"e6b311ab-c933-4950-9104-8071d75d77ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnjpj" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.258300 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/22ab7539-f66a-421e-8053-1cd8e03d465f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-7nk8v\" (UID: \"22ab7539-f66a-421e-8053-1cd8e03d465f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7nk8v" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.258362 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6b6ccb8c-d307-430b-9232-70497214c0bf-installation-pull-secrets\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.258411 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3c23d015-ef00-4a30-b2da-dc15d140a773-config-volume\") pod \"dns-default-d8x8n\" (UID: \"3c23d015-ef00-4a30-b2da-dc15d140a773\") " pod="openshift-dns/dns-default-d8x8n" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.258455 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-24f2z\" (UniqueName: \"kubernetes.io/projected/6b6ccb8c-d307-430b-9232-70497214c0bf-kube-api-access-24f2z\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.258471 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/bc099334-ba8a-4538-9a6d-98084ecb9a63-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-mhnj2\" (UID: \"bc099334-ba8a-4538-9a6d-98084ecb9a63\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mhnj2" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.258521 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22ab7539-f66a-421e-8053-1cd8e03d465f-config\") pod \"kube-controller-manager-operator-78b949d7b-7nk8v\" (UID: \"22ab7539-f66a-421e-8053-1cd8e03d465f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7nk8v" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.258550 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/e1a3f43f-63fe-4816-8414-2d7e4c63be57-mountpoint-dir\") pod \"csi-hostpathplugin-dnnwg\" (UID: \"e1a3f43f-63fe-4816-8414-2d7e4c63be57\") " pod="hostpath-provisioner/csi-hostpathplugin-dnnwg" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.258566 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/e1a3f43f-63fe-4816-8414-2d7e4c63be57-plugins-dir\") pod \"csi-hostpathplugin-dnnwg\" (UID: \"e1a3f43f-63fe-4816-8414-2d7e4c63be57\") " pod="hostpath-provisioner/csi-hostpathplugin-dnnwg" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.259192 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/e1a3f43f-63fe-4816-8414-2d7e4c63be57-csi-data-dir\") pod \"csi-hostpathplugin-dnnwg\" (UID: \"e1a3f43f-63fe-4816-8414-2d7e4c63be57\") " pod="hostpath-provisioner/csi-hostpathplugin-dnnwg" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.259236 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fkgdp\" (UniqueName: \"kubernetes.io/projected/e1a3f43f-63fe-4816-8414-2d7e4c63be57-kube-api-access-fkgdp\") pod \"csi-hostpathplugin-dnnwg\" (UID: \"e1a3f43f-63fe-4816-8414-2d7e4c63be57\") " pod="hostpath-provisioner/csi-hostpathplugin-dnnwg" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.259280 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b9509019-37bf-4674-af71-c259393eab0b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-49rxs\" (UID: \"b9509019-37bf-4674-af71-c259393eab0b\") " pod="openshift-marketplace/marketplace-operator-79b997595-49rxs" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.259318 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6b6ccb8c-d307-430b-9232-70497214c0bf-registry-tls\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.259335 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6b6ccb8c-d307-430b-9232-70497214c0bf-ca-trust-extracted\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.259376 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4z6rg\" (UniqueName: \"kubernetes.io/projected/bc099334-ba8a-4538-9a6d-98084ecb9a63-kube-api-access-4z6rg\") pod \"package-server-manager-789f6589d5-mhnj2\" (UID: \"bc099334-ba8a-4538-9a6d-98084ecb9a63\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mhnj2" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.259395 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-skjkj\" (UniqueName: \"kubernetes.io/projected/b9509019-37bf-4674-af71-c259393eab0b-kube-api-access-skjkj\") pod \"marketplace-operator-79b997595-49rxs\" (UID: \"b9509019-37bf-4674-af71-c259393eab0b\") " pod="openshift-marketplace/marketplace-operator-79b997595-49rxs" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.259527 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ee299c82-155e-4da8-9d33-80c85d5460d9-srv-cert\") pod \"catalog-operator-68c6474976-v7xvv\" (UID: \"ee299c82-155e-4da8-9d33-80c85d5460d9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v7xvv" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.259558 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6b6ccb8c-d307-430b-9232-70497214c0bf-registry-certificates\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.259619 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/e1a3f43f-63fe-4816-8414-2d7e4c63be57-registration-dir\") pod \"csi-hostpathplugin-dnnwg\" (UID: \"e1a3f43f-63fe-4816-8414-2d7e4c63be57\") " pod="hostpath-provisioner/csi-hostpathplugin-dnnwg" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.259651 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6b6ccb8c-d307-430b-9232-70497214c0bf-bound-sa-token\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.259669 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6b6ccb8c-d307-430b-9232-70497214c0bf-trusted-ca\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.263028 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mx5wl\" (UniqueName: \"kubernetes.io/projected/ee299c82-155e-4da8-9d33-80c85d5460d9-kube-api-access-mx5wl\") pod \"catalog-operator-68c6474976-v7xvv\" (UID: \"ee299c82-155e-4da8-9d33-80c85d5460d9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v7xvv" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.263122 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b9509019-37bf-4674-af71-c259393eab0b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-49rxs\" (UID: \"b9509019-37bf-4674-af71-c259393eab0b\") " pod="openshift-marketplace/marketplace-operator-79b997595-49rxs" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.265868 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.266316 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ee299c82-155e-4da8-9d33-80c85d5460d9-profile-collector-cert\") pod \"catalog-operator-68c6474976-v7xvv\" (UID: \"ee299c82-155e-4da8-9d33-80c85d5460d9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v7xvv" Oct 11 03:53:34 crc kubenswrapper[4967]: E1011 03:53:34.266590 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:34.766577357 +0000 UTC m=+142.729786290 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.278729 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3c23d015-ef00-4a30-b2da-dc15d140a773-metrics-tls\") pod \"dns-default-d8x8n\" (UID: \"3c23d015-ef00-4a30-b2da-dc15d140a773\") " pod="openshift-dns/dns-default-d8x8n" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.278947 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/de60a4e8-2c49-4b5b-a28d-b782c0852911-signing-key\") pod \"service-ca-9c57cc56f-4xkx2\" (UID: \"de60a4e8-2c49-4b5b-a28d-b782c0852911\") " pod="openshift-service-ca/service-ca-9c57cc56f-4xkx2" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.279082 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/315ce8c5-1d03-4b22-885d-04d07e83ab9d-serving-cert\") pod \"service-ca-operator-777779d784-9c49l\" (UID: \"315ce8c5-1d03-4b22-885d-04d07e83ab9d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-9c49l" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.279115 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/315ce8c5-1d03-4b22-885d-04d07e83ab9d-config\") pod \"service-ca-operator-777779d784-9c49l\" (UID: \"315ce8c5-1d03-4b22-885d-04d07e83ab9d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-9c49l" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.280635 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-g4q8m" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.280757 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rzd69\" (UniqueName: \"kubernetes.io/projected/315ce8c5-1d03-4b22-885d-04d07e83ab9d-kube-api-access-rzd69\") pod \"service-ca-operator-777779d784-9c49l\" (UID: \"315ce8c5-1d03-4b22-885d-04d07e83ab9d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-9c49l" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.281096 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/22ab7539-f66a-421e-8053-1cd8e03d465f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-7nk8v\" (UID: \"22ab7539-f66a-421e-8053-1cd8e03d465f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7nk8v" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.295503 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ws79r" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.295882 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-kdpld" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.326962 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-fsc8x" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.350586 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bjc5w"] Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.370695 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-csl6r"] Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.382492 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:34 crc kubenswrapper[4967]: E1011 03:53:34.382633 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:34.882612805 +0000 UTC m=+142.845821738 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.382731 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6b6ccb8c-d307-430b-9232-70497214c0bf-installation-pull-secrets\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.382757 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3c23d015-ef00-4a30-b2da-dc15d140a773-config-volume\") pod \"dns-default-d8x8n\" (UID: \"3c23d015-ef00-4a30-b2da-dc15d140a773\") " pod="openshift-dns/dns-default-d8x8n" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.382776 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6hft\" (UniqueName: \"kubernetes.io/projected/1b95aa73-24a0-496b-988c-5f0d52f1f00e-kube-api-access-s6hft\") pod \"ingress-canary-jg6t8\" (UID: \"1b95aa73-24a0-496b-988c-5f0d52f1f00e\") " pod="openshift-ingress-canary/ingress-canary-jg6t8" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.382801 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-24f2z\" (UniqueName: \"kubernetes.io/projected/6b6ccb8c-d307-430b-9232-70497214c0bf-kube-api-access-24f2z\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.382816 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/bc099334-ba8a-4538-9a6d-98084ecb9a63-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-mhnj2\" (UID: \"bc099334-ba8a-4538-9a6d-98084ecb9a63\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mhnj2" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.383615 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22ab7539-f66a-421e-8053-1cd8e03d465f-config\") pod \"kube-controller-manager-operator-78b949d7b-7nk8v\" (UID: \"22ab7539-f66a-421e-8053-1cd8e03d465f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7nk8v" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.383620 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/3c23d015-ef00-4a30-b2da-dc15d140a773-config-volume\") pod \"dns-default-d8x8n\" (UID: \"3c23d015-ef00-4a30-b2da-dc15d140a773\") " pod="openshift-dns/dns-default-d8x8n" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.383642 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/e1a3f43f-63fe-4816-8414-2d7e4c63be57-mountpoint-dir\") pod \"csi-hostpathplugin-dnnwg\" (UID: \"e1a3f43f-63fe-4816-8414-2d7e4c63be57\") " pod="hostpath-provisioner/csi-hostpathplugin-dnnwg" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.384085 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22ab7539-f66a-421e-8053-1cd8e03d465f-config\") pod \"kube-controller-manager-operator-78b949d7b-7nk8v\" (UID: \"22ab7539-f66a-421e-8053-1cd8e03d465f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7nk8v" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.384236 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/e1a3f43f-63fe-4816-8414-2d7e4c63be57-plugins-dir\") pod \"csi-hostpathplugin-dnnwg\" (UID: \"e1a3f43f-63fe-4816-8414-2d7e4c63be57\") " pod="hostpath-provisioner/csi-hostpathplugin-dnnwg" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.386804 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-cwz9p"] Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.387282 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6b6ccb8c-d307-430b-9232-70497214c0bf-installation-pull-secrets\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.388582 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/bc099334-ba8a-4538-9a6d-98084ecb9a63-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-mhnj2\" (UID: \"bc099334-ba8a-4538-9a6d-98084ecb9a63\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mhnj2" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.390933 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/e1a3f43f-63fe-4816-8414-2d7e4c63be57-mountpoint-dir\") pod \"csi-hostpathplugin-dnnwg\" (UID: \"e1a3f43f-63fe-4816-8414-2d7e4c63be57\") " pod="hostpath-provisioner/csi-hostpathplugin-dnnwg" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.391042 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/e1a3f43f-63fe-4816-8414-2d7e4c63be57-plugins-dir\") pod \"csi-hostpathplugin-dnnwg\" (UID: \"e1a3f43f-63fe-4816-8414-2d7e4c63be57\") " pod="hostpath-provisioner/csi-hostpathplugin-dnnwg" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.393063 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/e1a3f43f-63fe-4816-8414-2d7e4c63be57-csi-data-dir\") pod \"csi-hostpathplugin-dnnwg\" (UID: \"e1a3f43f-63fe-4816-8414-2d7e4c63be57\") " pod="hostpath-provisioner/csi-hostpathplugin-dnnwg" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.393122 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fkgdp\" (UniqueName: \"kubernetes.io/projected/e1a3f43f-63fe-4816-8414-2d7e4c63be57-kube-api-access-fkgdp\") pod \"csi-hostpathplugin-dnnwg\" (UID: \"e1a3f43f-63fe-4816-8414-2d7e4c63be57\") " pod="hostpath-provisioner/csi-hostpathplugin-dnnwg" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.393144 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b9509019-37bf-4674-af71-c259393eab0b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-49rxs\" (UID: \"b9509019-37bf-4674-af71-c259393eab0b\") " pod="openshift-marketplace/marketplace-operator-79b997595-49rxs" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.393162 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6b6ccb8c-d307-430b-9232-70497214c0bf-registry-tls\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.393277 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6b6ccb8c-d307-430b-9232-70497214c0bf-ca-trust-extracted\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.393305 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4z6rg\" (UniqueName: \"kubernetes.io/projected/bc099334-ba8a-4538-9a6d-98084ecb9a63-kube-api-access-4z6rg\") pod \"package-server-manager-789f6589d5-mhnj2\" (UID: \"bc099334-ba8a-4538-9a6d-98084ecb9a63\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mhnj2" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.393321 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-skjkj\" (UniqueName: \"kubernetes.io/projected/b9509019-37bf-4674-af71-c259393eab0b-kube-api-access-skjkj\") pod \"marketplace-operator-79b997595-49rxs\" (UID: \"b9509019-37bf-4674-af71-c259393eab0b\") " pod="openshift-marketplace/marketplace-operator-79b997595-49rxs" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.393339 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ee299c82-155e-4da8-9d33-80c85d5460d9-srv-cert\") pod \"catalog-operator-68c6474976-v7xvv\" (UID: \"ee299c82-155e-4da8-9d33-80c85d5460d9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v7xvv" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.393357 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6b6ccb8c-d307-430b-9232-70497214c0bf-registry-certificates\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.393394 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/e1a3f43f-63fe-4816-8414-2d7e4c63be57-registration-dir\") pod \"csi-hostpathplugin-dnnwg\" (UID: \"e1a3f43f-63fe-4816-8414-2d7e4c63be57\") " pod="hostpath-provisioner/csi-hostpathplugin-dnnwg" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.394099 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6b6ccb8c-d307-430b-9232-70497214c0bf-bound-sa-token\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.394124 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1b95aa73-24a0-496b-988c-5f0d52f1f00e-cert\") pod \"ingress-canary-jg6t8\" (UID: \"1b95aa73-24a0-496b-988c-5f0d52f1f00e\") " pod="openshift-ingress-canary/ingress-canary-jg6t8" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.394150 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6b6ccb8c-d307-430b-9232-70497214c0bf-trusted-ca\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.394174 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mx5wl\" (UniqueName: \"kubernetes.io/projected/ee299c82-155e-4da8-9d33-80c85d5460d9-kube-api-access-mx5wl\") pod \"catalog-operator-68c6474976-v7xvv\" (UID: \"ee299c82-155e-4da8-9d33-80c85d5460d9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v7xvv" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.394205 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b9509019-37bf-4674-af71-c259393eab0b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-49rxs\" (UID: \"b9509019-37bf-4674-af71-c259393eab0b\") " pod="openshift-marketplace/marketplace-operator-79b997595-49rxs" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.394247 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.394272 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ee299c82-155e-4da8-9d33-80c85d5460d9-profile-collector-cert\") pod \"catalog-operator-68c6474976-v7xvv\" (UID: \"ee299c82-155e-4da8-9d33-80c85d5460d9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v7xvv" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.394302 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3c23d015-ef00-4a30-b2da-dc15d140a773-metrics-tls\") pod \"dns-default-d8x8n\" (UID: \"3c23d015-ef00-4a30-b2da-dc15d140a773\") " pod="openshift-dns/dns-default-d8x8n" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.394362 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/de60a4e8-2c49-4b5b-a28d-b782c0852911-signing-key\") pod \"service-ca-9c57cc56f-4xkx2\" (UID: \"de60a4e8-2c49-4b5b-a28d-b782c0852911\") " pod="openshift-service-ca/service-ca-9c57cc56f-4xkx2" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.394381 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/315ce8c5-1d03-4b22-885d-04d07e83ab9d-serving-cert\") pod \"service-ca-operator-777779d784-9c49l\" (UID: \"315ce8c5-1d03-4b22-885d-04d07e83ab9d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-9c49l" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.394395 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/315ce8c5-1d03-4b22-885d-04d07e83ab9d-config\") pod \"service-ca-operator-777779d784-9c49l\" (UID: \"315ce8c5-1d03-4b22-885d-04d07e83ab9d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-9c49l" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.394410 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rzd69\" (UniqueName: \"kubernetes.io/projected/315ce8c5-1d03-4b22-885d-04d07e83ab9d-kube-api-access-rzd69\") pod \"service-ca-operator-777779d784-9c49l\" (UID: \"315ce8c5-1d03-4b22-885d-04d07e83ab9d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-9c49l" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.394431 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/22ab7539-f66a-421e-8053-1cd8e03d465f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-7nk8v\" (UID: \"22ab7539-f66a-421e-8053-1cd8e03d465f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7nk8v" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.394469 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n5dnf\" (UniqueName: \"kubernetes.io/projected/e6b311ab-c933-4950-9104-8071d75d77ac-kube-api-access-n5dnf\") pod \"packageserver-d55dfcdfc-wnjpj\" (UID: \"e6b311ab-c933-4950-9104-8071d75d77ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnjpj" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.394494 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sb5vr\" (UniqueName: \"kubernetes.io/projected/3c23d015-ef00-4a30-b2da-dc15d140a773-kube-api-access-sb5vr\") pod \"dns-default-d8x8n\" (UID: \"3c23d015-ef00-4a30-b2da-dc15d140a773\") " pod="openshift-dns/dns-default-d8x8n" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.394512 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w7w9f\" (UniqueName: \"kubernetes.io/projected/de60a4e8-2c49-4b5b-a28d-b782c0852911-kube-api-access-w7w9f\") pod \"service-ca-9c57cc56f-4xkx2\" (UID: \"de60a4e8-2c49-4b5b-a28d-b782c0852911\") " pod="openshift-service-ca/service-ca-9c57cc56f-4xkx2" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.396116 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e6b311ab-c933-4950-9104-8071d75d77ac-apiservice-cert\") pod \"packageserver-d55dfcdfc-wnjpj\" (UID: \"e6b311ab-c933-4950-9104-8071d75d77ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnjpj" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.396146 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/de60a4e8-2c49-4b5b-a28d-b782c0852911-signing-cabundle\") pod \"service-ca-9c57cc56f-4xkx2\" (UID: \"de60a4e8-2c49-4b5b-a28d-b782c0852911\") " pod="openshift-service-ca/service-ca-9c57cc56f-4xkx2" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.396253 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/e1a3f43f-63fe-4816-8414-2d7e4c63be57-socket-dir\") pod \"csi-hostpathplugin-dnnwg\" (UID: \"e1a3f43f-63fe-4816-8414-2d7e4c63be57\") " pod="hostpath-provisioner/csi-hostpathplugin-dnnwg" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.396269 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e6b311ab-c933-4950-9104-8071d75d77ac-webhook-cert\") pod \"packageserver-d55dfcdfc-wnjpj\" (UID: \"e6b311ab-c933-4950-9104-8071d75d77ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnjpj" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.396297 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/e6b311ab-c933-4950-9104-8071d75d77ac-tmpfs\") pod \"packageserver-d55dfcdfc-wnjpj\" (UID: \"e6b311ab-c933-4950-9104-8071d75d77ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnjpj" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.396490 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/22ab7539-f66a-421e-8053-1cd8e03d465f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-7nk8v\" (UID: \"22ab7539-f66a-421e-8053-1cd8e03d465f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7nk8v" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.400213 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6b6ccb8c-d307-430b-9232-70497214c0bf-registry-tls\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.393391 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/e1a3f43f-63fe-4816-8414-2d7e4c63be57-csi-data-dir\") pod \"csi-hostpathplugin-dnnwg\" (UID: \"e1a3f43f-63fe-4816-8414-2d7e4c63be57\") " pod="hostpath-provisioner/csi-hostpathplugin-dnnwg" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.401037 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6b6ccb8c-d307-430b-9232-70497214c0bf-registry-certificates\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.401268 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/e1a3f43f-63fe-4816-8414-2d7e4c63be57-registration-dir\") pod \"csi-hostpathplugin-dnnwg\" (UID: \"e1a3f43f-63fe-4816-8414-2d7e4c63be57\") " pod="hostpath-provisioner/csi-hostpathplugin-dnnwg" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.402761 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b9509019-37bf-4674-af71-c259393eab0b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-49rxs\" (UID: \"b9509019-37bf-4674-af71-c259393eab0b\") " pod="openshift-marketplace/marketplace-operator-79b997595-49rxs" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.403933 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/de60a4e8-2c49-4b5b-a28d-b782c0852911-signing-key\") pod \"service-ca-9c57cc56f-4xkx2\" (UID: \"de60a4e8-2c49-4b5b-a28d-b782c0852911\") " pod="openshift-service-ca/service-ca-9c57cc56f-4xkx2" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.405717 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/de60a4e8-2c49-4b5b-a28d-b782c0852911-signing-cabundle\") pod \"service-ca-9c57cc56f-4xkx2\" (UID: \"de60a4e8-2c49-4b5b-a28d-b782c0852911\") " pod="openshift-service-ca/service-ca-9c57cc56f-4xkx2" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.405873 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e6b311ab-c933-4950-9104-8071d75d77ac-apiservice-cert\") pod \"packageserver-d55dfcdfc-wnjpj\" (UID: \"e6b311ab-c933-4950-9104-8071d75d77ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnjpj" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.405973 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6b6ccb8c-d307-430b-9232-70497214c0bf-ca-trust-extracted\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.406372 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/e6b311ab-c933-4950-9104-8071d75d77ac-tmpfs\") pod \"packageserver-d55dfcdfc-wnjpj\" (UID: \"e6b311ab-c933-4950-9104-8071d75d77ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnjpj" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.407132 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/315ce8c5-1d03-4b22-885d-04d07e83ab9d-config\") pod \"service-ca-operator-777779d784-9c49l\" (UID: \"315ce8c5-1d03-4b22-885d-04d07e83ab9d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-9c49l" Oct 11 03:53:34 crc kubenswrapper[4967]: E1011 03:53:34.407696 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:34.907684239 +0000 UTC m=+142.870893172 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.408776 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/22ab7539-f66a-421e-8053-1cd8e03d465f-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-7nk8v\" (UID: \"22ab7539-f66a-421e-8053-1cd8e03d465f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7nk8v" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.409870 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/ee299c82-155e-4da8-9d33-80c85d5460d9-profile-collector-cert\") pod \"catalog-operator-68c6474976-v7xvv\" (UID: \"ee299c82-155e-4da8-9d33-80c85d5460d9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v7xvv" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.409942 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/e1a3f43f-63fe-4816-8414-2d7e4c63be57-socket-dir\") pod \"csi-hostpathplugin-dnnwg\" (UID: \"e1a3f43f-63fe-4816-8414-2d7e4c63be57\") " pod="hostpath-provisioner/csi-hostpathplugin-dnnwg" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.411377 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b9509019-37bf-4674-af71-c259393eab0b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-49rxs\" (UID: \"b9509019-37bf-4674-af71-c259393eab0b\") " pod="openshift-marketplace/marketplace-operator-79b997595-49rxs" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.411361 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/315ce8c5-1d03-4b22-885d-04d07e83ab9d-serving-cert\") pod \"service-ca-operator-777779d784-9c49l\" (UID: \"315ce8c5-1d03-4b22-885d-04d07e83ab9d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-9c49l" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.411804 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6b6ccb8c-d307-430b-9232-70497214c0bf-trusted-ca\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.412156 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e6b311ab-c933-4950-9104-8071d75d77ac-webhook-cert\") pod \"packageserver-d55dfcdfc-wnjpj\" (UID: \"e6b311ab-c933-4950-9104-8071d75d77ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnjpj" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.416098 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/3c23d015-ef00-4a30-b2da-dc15d140a773-metrics-tls\") pod \"dns-default-d8x8n\" (UID: \"3c23d015-ef00-4a30-b2da-dc15d140a773\") " pod="openshift-dns/dns-default-d8x8n" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.416395 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/ee299c82-155e-4da8-9d33-80c85d5460d9-srv-cert\") pod \"catalog-operator-68c6474976-v7xvv\" (UID: \"ee299c82-155e-4da8-9d33-80c85d5460d9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v7xvv" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.428847 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bxjls"] Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.430108 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-24f2z\" (UniqueName: \"kubernetes.io/projected/6b6ccb8c-d307-430b-9232-70497214c0bf-kube-api-access-24f2z\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.439427 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gwrxz"] Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.447896 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4z6rg\" (UniqueName: \"kubernetes.io/projected/bc099334-ba8a-4538-9a6d-98084ecb9a63-kube-api-access-4z6rg\") pod \"package-server-manager-789f6589d5-mhnj2\" (UID: \"bc099334-ba8a-4538-9a6d-98084ecb9a63\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mhnj2" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.451749 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rmkqj"] Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.451786 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kmpsc"] Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.452719 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-7sxxj"] Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.471762 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fkgdp\" (UniqueName: \"kubernetes.io/projected/e1a3f43f-63fe-4816-8414-2d7e4c63be57-kube-api-access-fkgdp\") pod \"csi-hostpathplugin-dnnwg\" (UID: \"e1a3f43f-63fe-4816-8414-2d7e4c63be57\") " pod="hostpath-provisioner/csi-hostpathplugin-dnnwg" Oct 11 03:53:34 crc kubenswrapper[4967]: W1011 03:53:34.492882 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5e20a512_8f34_4d21_be45_e749f7a6cc87.slice/crio-34766231a2f79dc50e67406c04d6e72ed0012f85c3c3e3ca47d3ff1e82eaa613 WatchSource:0}: Error finding container 34766231a2f79dc50e67406c04d6e72ed0012f85c3c3e3ca47d3ff1e82eaa613: Status 404 returned error can't find the container with id 34766231a2f79dc50e67406c04d6e72ed0012f85c3c3e3ca47d3ff1e82eaa613 Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.498034 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.498228 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6hft\" (UniqueName: \"kubernetes.io/projected/1b95aa73-24a0-496b-988c-5f0d52f1f00e-kube-api-access-s6hft\") pod \"ingress-canary-jg6t8\" (UID: \"1b95aa73-24a0-496b-988c-5f0d52f1f00e\") " pod="openshift-ingress-canary/ingress-canary-jg6t8" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.498284 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1b95aa73-24a0-496b-988c-5f0d52f1f00e-cert\") pod \"ingress-canary-jg6t8\" (UID: \"1b95aa73-24a0-496b-988c-5f0d52f1f00e\") " pod="openshift-ingress-canary/ingress-canary-jg6t8" Oct 11 03:53:34 crc kubenswrapper[4967]: E1011 03:53:34.498570 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:34.998553659 +0000 UTC m=+142.961762592 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.501477 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/1b95aa73-24a0-496b-988c-5f0d52f1f00e-cert\") pod \"ingress-canary-jg6t8\" (UID: \"1b95aa73-24a0-496b-988c-5f0d52f1f00e\") " pod="openshift-ingress-canary/ingress-canary-jg6t8" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.501567 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-skjkj\" (UniqueName: \"kubernetes.io/projected/b9509019-37bf-4674-af71-c259393eab0b-kube-api-access-skjkj\") pod \"marketplace-operator-79b997595-49rxs\" (UID: \"b9509019-37bf-4674-af71-c259393eab0b\") " pod="openshift-marketplace/marketplace-operator-79b997595-49rxs" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.516854 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-ws6r9"] Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.520673 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rzd69\" (UniqueName: \"kubernetes.io/projected/315ce8c5-1d03-4b22-885d-04d07e83ab9d-kube-api-access-rzd69\") pod \"service-ca-operator-777779d784-9c49l\" (UID: \"315ce8c5-1d03-4b22-885d-04d07e83ab9d\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-9c49l" Oct 11 03:53:34 crc kubenswrapper[4967]: W1011 03:53:34.523695 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod49347426_bca5_4fde_8b87_21dc8fc486d8.slice/crio-d82ae857478d93f4819301b0fdcfeb1789a43ca664a487510e257b538aa7d67f WatchSource:0}: Error finding container d82ae857478d93f4819301b0fdcfeb1789a43ca664a487510e257b538aa7d67f: Status 404 returned error can't find the container with id d82ae857478d93f4819301b0fdcfeb1789a43ca664a487510e257b538aa7d67f Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.527551 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tr7z7"] Oct 11 03:53:34 crc kubenswrapper[4967]: W1011 03:53:34.527662 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3883197f_52dd_465a_82bb_c4dd32b3d090.slice/crio-19723a34c93da8c18c43d9f9130ec09503d0cad8d9f120df78b56981c442f781 WatchSource:0}: Error finding container 19723a34c93da8c18c43d9f9130ec09503d0cad8d9f120df78b56981c442f781: Status 404 returned error can't find the container with id 19723a34c93da8c18c43d9f9130ec09503d0cad8d9f120df78b56981c442f781 Oct 11 03:53:34 crc kubenswrapper[4967]: W1011 03:53:34.529394 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod45ebd02f_dc19_48a5_8309_bdbf11a8a7a4.slice/crio-2ecba630a9c447b1e1a55d6eee6d9074d8ef56c50bac91a9aaa8ca1ff569443f WatchSource:0}: Error finding container 2ecba630a9c447b1e1a55d6eee6d9074d8ef56c50bac91a9aaa8ca1ff569443f: Status 404 returned error can't find the container with id 2ecba630a9c447b1e1a55d6eee6d9074d8ef56c50bac91a9aaa8ca1ff569443f Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.529579 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-dtj88"] Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.544963 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/22ab7539-f66a-421e-8053-1cd8e03d465f-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-7nk8v\" (UID: \"22ab7539-f66a-421e-8053-1cd8e03d465f\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7nk8v" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.551776 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-n6qrf"] Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.561909 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6b6ccb8c-d307-430b-9232-70497214c0bf-bound-sa-token\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:34 crc kubenswrapper[4967]: W1011 03:53:34.574541 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod092da80d_ee9c_4e88_b3e2_58bb3dd77f9b.slice/crio-8b84d45c585d6f8c52d51403a21febb8fef40eb2a08792f5662f04c22a7fb31b WatchSource:0}: Error finding container 8b84d45c585d6f8c52d51403a21febb8fef40eb2a08792f5662f04c22a7fb31b: Status 404 returned error can't find the container with id 8b84d45c585d6f8c52d51403a21febb8fef40eb2a08792f5662f04c22a7fb31b Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.575321 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mhnj2" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.576718 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gwrxz" event={"ID":"ac8fc0d4-aa55-4974-a506-7ccd2713f31a","Type":"ContainerStarted","Data":"5f55544214c7b9cbf7b409b33e3a15920e2db60dfaf3316f0b0bcfba8bc1ca96"} Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.581285 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-49rxs" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.581878 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sb5vr\" (UniqueName: \"kubernetes.io/projected/3c23d015-ef00-4a30-b2da-dc15d140a773-kube-api-access-sb5vr\") pod \"dns-default-d8x8n\" (UID: \"3c23d015-ef00-4a30-b2da-dc15d140a773\") " pod="openshift-dns/dns-default-d8x8n" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.590430 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-9c49l" Oct 11 03:53:34 crc kubenswrapper[4967]: W1011 03:53:34.591954 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd6b9a450_fd86_47a3_9c0e_f61aa7a24377.slice/crio-9e23760ef5636af900f586efabf514fdb67a1b319c96e4e989f818533ca865cc WatchSource:0}: Error finding container 9e23760ef5636af900f586efabf514fdb67a1b319c96e4e989f818533ca865cc: Status 404 returned error can't find the container with id 9e23760ef5636af900f586efabf514fdb67a1b319c96e4e989f818533ca865cc Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.595226 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-92sz4" event={"ID":"1c77fe9a-d2e1-465b-8600-9f7f2d1fa206","Type":"ContainerStarted","Data":"f8cb425ed8afecfff544a72cae16fd56f27a0a17fdf2156e2e4dae6ab1ba7fc4"} Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.595264 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-92sz4" event={"ID":"1c77fe9a-d2e1-465b-8600-9f7f2d1fa206","Type":"ContainerStarted","Data":"8675633416cc7576fe45dc4505545a7520fc1778845f60d19e362df7a3e1480a"} Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.596384 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-92sz4" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.597934 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rw8ng"] Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.599227 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:34 crc kubenswrapper[4967]: E1011 03:53:34.600916 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:35.100901084 +0000 UTC m=+143.064110017 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.608234 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-qn4g8" event={"ID":"3bc0bb1d-5b22-4e74-8aa8-6554b77a2b73","Type":"ContainerStarted","Data":"8dce7aea7026cdcdb2c0ec193e2522e5ce2350a06087e7b841754d598cd7c9e6"} Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.608273 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-qn4g8" event={"ID":"3bc0bb1d-5b22-4e74-8aa8-6554b77a2b73","Type":"ContainerStarted","Data":"fd9240229669708d386560b69580e0c1d190c49fa47a1895a64e32d0f5381ae9"} Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.608466 4967 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-92sz4 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" start-of-body= Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.608556 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-92sz4" podUID="1c77fe9a-d2e1-465b-8600-9f7f2d1fa206" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.5:8443/healthz\": dial tcp 10.217.0.5:8443: connect: connection refused" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.613234 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n5dnf\" (UniqueName: \"kubernetes.io/projected/e6b311ab-c933-4950-9104-8071d75d77ac-kube-api-access-n5dnf\") pod \"packageserver-d55dfcdfc-wnjpj\" (UID: \"e6b311ab-c933-4950-9104-8071d75d77ac\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnjpj" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.619520 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-dnnwg" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.621705 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w7w9f\" (UniqueName: \"kubernetes.io/projected/de60a4e8-2c49-4b5b-a28d-b782c0852911-kube-api-access-w7w9f\") pod \"service-ca-9c57cc56f-4xkx2\" (UID: \"de60a4e8-2c49-4b5b-a28d-b782c0852911\") " pod="openshift-service-ca/service-ca-9c57cc56f-4xkx2" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.634897 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-d8x8n" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.638981 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mx5wl\" (UniqueName: \"kubernetes.io/projected/ee299c82-155e-4da8-9d33-80c85d5460d9-kube-api-access-mx5wl\") pod \"catalog-operator-68c6474976-v7xvv\" (UID: \"ee299c82-155e-4da8-9d33-80c85d5460d9\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v7xvv" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.650999 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-rt5vd" event={"ID":"a4ef12d4-a717-4afa-b3c7-179ea0f7f16e","Type":"ContainerStarted","Data":"61fd90ffd53cf50b420a7553b2e4525ffb78504c7c0e14db663ea057e73d4dc0"} Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.651040 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-rt5vd" event={"ID":"a4ef12d4-a717-4afa-b3c7-179ea0f7f16e","Type":"ContainerStarted","Data":"72a90e18a41753bb9c46909491a93c5f6392facd793acf6b50939acc7d2d48c8"} Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.651447 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-rt5vd" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.657938 4967 patch_prober.go:28] interesting pod/console-operator-58897d9998-rt5vd container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" start-of-body= Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.657985 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-rt5vd" podUID="a4ef12d4-a717-4afa-b3c7-179ea0f7f16e" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.11:8443/readyz\": dial tcp 10.217.0.11:8443: connect: connection refused" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.687927 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6hft\" (UniqueName: \"kubernetes.io/projected/1b95aa73-24a0-496b-988c-5f0d52f1f00e-kube-api-access-s6hft\") pod \"ingress-canary-jg6t8\" (UID: \"1b95aa73-24a0-496b-988c-5f0d52f1f00e\") " pod="openshift-ingress-canary/ingress-canary-jg6t8" Oct 11 03:53:34 crc kubenswrapper[4967]: W1011 03:53:34.688836 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0761866e_d6ab_40c9_a776_8c7fb2683fad.slice/crio-3a1a4bbfa346fdb02307a46cef0ee3f6ff8eb1b049cfa6b21075a19211cbca38 WatchSource:0}: Error finding container 3a1a4bbfa346fdb02307a46cef0ee3f6ff8eb1b049cfa6b21075a19211cbca38: Status 404 returned error can't find the container with id 3a1a4bbfa346fdb02307a46cef0ee3f6ff8eb1b049cfa6b21075a19211cbca38 Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.699587 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:34 crc kubenswrapper[4967]: E1011 03:53:34.700648 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:35.200631766 +0000 UTC m=+143.163840699 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.724855 4967 generic.go:334] "Generic (PLEG): container finished" podID="454804e7-0d4d-4536-8544-87a5b6368f70" containerID="b7f528ca66861215d4a00fef6b1388aa43f683cb4222399ce4513ad40f01a346" exitCode=0 Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.724918 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" event={"ID":"454804e7-0d4d-4536-8544-87a5b6368f70","Type":"ContainerDied","Data":"b7f528ca66861215d4a00fef6b1388aa43f683cb4222399ce4513ad40f01a346"} Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.724955 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" event={"ID":"454804e7-0d4d-4536-8544-87a5b6368f70","Type":"ContainerStarted","Data":"a3e660d8a95d16aea11ecefb9f58831c58e187a75b3c57a6ec27e614f3dfc49d"} Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.732282 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bxjls" event={"ID":"89f0e615-1fde-483d-9b38-83287f104360","Type":"ContainerStarted","Data":"93a78fa9a27ed85d1d3421d4bfa8857a3e80d2036ae55191284a7db2211ea9c8"} Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.733591 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-csl6r" event={"ID":"5e20a512-8f34-4d21-be45-e749f7a6cc87","Type":"ContainerStarted","Data":"34766231a2f79dc50e67406c04d6e72ed0012f85c3c3e3ca47d3ff1e82eaa613"} Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.735645 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cwz9p" event={"ID":"038209e7-3fdb-490b-976e-f1791bff29d8","Type":"ContainerStarted","Data":"86a5199417e11b24a4d563ce1f913668edc4e9d6dba08425cef626e94dbe5d12"} Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.745354 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kmpsc" event={"ID":"49347426-bca5-4fde-8b87-21dc8fc486d8","Type":"ContainerStarted","Data":"d82ae857478d93f4819301b0fdcfeb1789a43ca664a487510e257b538aa7d67f"} Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.746507 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bjc5w" event={"ID":"8f4cc87b-f2fd-4c61-82fb-f362b81445d7","Type":"ContainerStarted","Data":"f262dd4c191c452c356e7fa6f91c4bbdc2df500650852d756537d2f7446b6385"} Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.758387 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" event={"ID":"0dae89fc-445f-4531-8928-29b334636522","Type":"ContainerStarted","Data":"81dc34607ed3916b6a34b0cad97035721d5397b349c54feaf67139e5c9de87e8"} Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.770840 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rv6n2"] Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.772056 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwp4c" event={"ID":"943c9c2f-24f5-408c-8e8d-c344253fc8f6","Type":"ContainerStarted","Data":"c9282381a84cc61757688abb2ade9c29509c46f20867f8f12decebddc2545444"} Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.787501 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-7sxxj" event={"ID":"45ebd02f-dc19-48a5-8309-bdbf11a8a7a4","Type":"ContainerStarted","Data":"2ecba630a9c447b1e1a55d6eee6d9074d8ef56c50bac91a9aaa8ca1ff569443f"} Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.788464 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rmkqj" event={"ID":"3883197f-52dd-465a-82bb-c4dd32b3d090","Type":"ContainerStarted","Data":"19723a34c93da8c18c43d9f9130ec09503d0cad8d9f120df78b56981c442f781"} Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.801102 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:34 crc kubenswrapper[4967]: E1011 03:53:34.801466 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:35.301452918 +0000 UTC m=+143.264661851 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.814812 4967 generic.go:334] "Generic (PLEG): container finished" podID="efe9c94d-8f97-4274-a494-a5ecf5b88cec" containerID="fb717c6a3971cbf20ff63d368ec4aedf6a7fc6577a7db690f7d7a82b5f209b55" exitCode=0 Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.826554 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7nk8v" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.830926 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" event={"ID":"efe9c94d-8f97-4274-a494-a5ecf5b88cec","Type":"ContainerDied","Data":"fb717c6a3971cbf20ff63d368ec4aedf6a7fc6577a7db690f7d7a82b5f209b55"} Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.831042 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" event={"ID":"efe9c94d-8f97-4274-a494-a5ecf5b88cec","Type":"ContainerStarted","Data":"f13d1c29f183e0cd49397d70d281370502eab2716bef2c7367f7bb83d71ebc91"} Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.838689 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v7xvv" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.842254 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnjpj" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.846213 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-mrsjk" event={"ID":"9b3efa80-8370-4353-b688-4701c4deec8f","Type":"ContainerStarted","Data":"453e859485f53221c31929592efdc31f2cf0522103706e058f48c7b7529bd0c3"} Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.846360 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-mrsjk" event={"ID":"9b3efa80-8370-4353-b688-4701c4deec8f","Type":"ContainerStarted","Data":"176f30a1fce96ca491b430a3eee8a3b89a5aa509b1f966aa0d21208d0fc99762"} Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.846448 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-mrsjk" event={"ID":"9b3efa80-8370-4353-b688-4701c4deec8f","Type":"ContainerStarted","Data":"ac0ddf6d1149c137adcf03a36d895900d3dbb86b6c7d825bb209d1172d681456"} Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.866884 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-9jjwn" event={"ID":"b5fca2ad-c70a-43e2-aff1-f180617a94f8","Type":"ContainerStarted","Data":"f8a91aa92c40c5527fb2f37654d1107888983f199b16a060acde35cb3e736d2b"} Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.882644 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-cp2n5"] Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.903715 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-4xkx2" Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.904030 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:34 crc kubenswrapper[4967]: E1011 03:53:34.904311 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:35.404290865 +0000 UTC m=+143.367499798 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.904497 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:34 crc kubenswrapper[4967]: E1011 03:53:34.905829 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:35.405818707 +0000 UTC m=+143.369027760 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.928865 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wtn79"] Oct 11 03:53:34 crc kubenswrapper[4967]: I1011 03:53:34.941000 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-jg6t8" Oct 11 03:53:35 crc kubenswrapper[4967]: I1011 03:53:35.011518 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:35 crc kubenswrapper[4967]: E1011 03:53:35.011658 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:35.511632355 +0000 UTC m=+143.474841288 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:35 crc kubenswrapper[4967]: I1011 03:53:35.012039 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:35 crc kubenswrapper[4967]: E1011 03:53:35.012360 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:35.512346434 +0000 UTC m=+143.475555367 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:35 crc kubenswrapper[4967]: I1011 03:53:35.062125 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-92sz4" podStartSLOduration=122.062107283 podStartE2EDuration="2m2.062107283s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:35.06050332 +0000 UTC m=+143.023712253" watchObservedRunningTime="2025-10-11 03:53:35.062107283 +0000 UTC m=+143.025316216" Oct 11 03:53:35 crc kubenswrapper[4967]: I1011 03:53:35.113190 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:35 crc kubenswrapper[4967]: E1011 03:53:35.113505 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:35.613487206 +0000 UTC m=+143.576696139 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:35 crc kubenswrapper[4967]: I1011 03:53:35.161005 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-gbqfq"] Oct 11 03:53:35 crc kubenswrapper[4967]: I1011 03:53:35.215981 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:35 crc kubenswrapper[4967]: E1011 03:53:35.216347 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:35.716335783 +0000 UTC m=+143.679544706 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:35 crc kubenswrapper[4967]: I1011 03:53:35.260412 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-rt5vd" podStartSLOduration=122.260396197 podStartE2EDuration="2m2.260396197s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:35.257821816 +0000 UTC m=+143.221030739" watchObservedRunningTime="2025-10-11 03:53:35.260396197 +0000 UTC m=+143.223605130" Oct 11 03:53:35 crc kubenswrapper[4967]: I1011 03:53:35.319488 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:35 crc kubenswrapper[4967]: E1011 03:53:35.324457 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:35.824434614 +0000 UTC m=+143.787643547 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:35 crc kubenswrapper[4967]: I1011 03:53:35.389108 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mhnj2"] Oct 11 03:53:35 crc kubenswrapper[4967]: I1011 03:53:35.426660 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:35 crc kubenswrapper[4967]: E1011 03:53:35.426985 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:35.926973974 +0000 UTC m=+143.890182907 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:35 crc kubenswrapper[4967]: I1011 03:53:35.435481 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-qn4g8" podStartSLOduration=122.435466046 podStartE2EDuration="2m2.435466046s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:35.433145952 +0000 UTC m=+143.396354885" watchObservedRunningTime="2025-10-11 03:53:35.435466046 +0000 UTC m=+143.398674979" Oct 11 03:53:35 crc kubenswrapper[4967]: I1011 03:53:35.468165 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ws79r"] Oct 11 03:53:35 crc kubenswrapper[4967]: I1011 03:53:35.492344 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-g4q8m"] Oct 11 03:53:35 crc kubenswrapper[4967]: I1011 03:53:35.496020 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-9c49l"] Oct 11 03:53:35 crc kubenswrapper[4967]: I1011 03:53:35.498150 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-49rxs"] Oct 11 03:53:35 crc kubenswrapper[4967]: I1011 03:53:35.528122 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:35 crc kubenswrapper[4967]: E1011 03:53:35.528577 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:36.028559986 +0000 UTC m=+143.991768909 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:35 crc kubenswrapper[4967]: I1011 03:53:35.528722 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:35 crc kubenswrapper[4967]: E1011 03:53:35.529099 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:36.029091331 +0000 UTC m=+143.992300264 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:35 crc kubenswrapper[4967]: I1011 03:53:35.557401 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-d8x8n"] Oct 11 03:53:35 crc kubenswrapper[4967]: I1011 03:53:35.630216 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:35 crc kubenswrapper[4967]: E1011 03:53:35.630597 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:36.130579882 +0000 UTC m=+144.093788815 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:35 crc kubenswrapper[4967]: I1011 03:53:35.667362 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335905-kdpld"] Oct 11 03:53:35 crc kubenswrapper[4967]: I1011 03:53:35.749474 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:35 crc kubenswrapper[4967]: E1011 03:53:35.749774 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:36.249762345 +0000 UTC m=+144.212971278 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:35 crc kubenswrapper[4967]: I1011 03:53:35.804202 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7nk8v"] Oct 11 03:53:35 crc kubenswrapper[4967]: I1011 03:53:35.852102 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:35 crc kubenswrapper[4967]: E1011 03:53:35.852392 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:36.352367616 +0000 UTC m=+144.315576549 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:35 crc kubenswrapper[4967]: I1011 03:53:35.893710 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-dnnwg"] Oct 11 03:53:35 crc kubenswrapper[4967]: I1011 03:53:35.929263 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wtn79" event={"ID":"5449c948-487e-426d-b151-31e8079576f3","Type":"ContainerStarted","Data":"f540a51e3cfc72196043bc956e60b32300dabd6c5ae10b9d0a1dd1a07b86aff6"} Oct 11 03:53:35 crc kubenswrapper[4967]: I1011 03:53:35.955389 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:35 crc kubenswrapper[4967]: E1011 03:53:35.955707 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:36.455695916 +0000 UTC m=+144.418904849 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:35 crc kubenswrapper[4967]: I1011 03:53:35.957876 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwp4c" event={"ID":"943c9c2f-24f5-408c-8e8d-c344253fc8f6","Type":"ContainerStarted","Data":"160d0d7e6bcb01899ddc54964316b15d14ac4566473633b1998dedf12916266f"} Oct 11 03:53:35 crc kubenswrapper[4967]: I1011 03:53:35.961191 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rw8ng" event={"ID":"b173cdfc-6a31-4d2e-93cd-a04eb1c68877","Type":"ContainerStarted","Data":"8d3f3e5e95b08609bb65bf338447d3f0137f003c6b2e19ae2b39651b4508034f"} Oct 11 03:53:35 crc kubenswrapper[4967]: I1011 03:53:35.963575 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-d8x8n" event={"ID":"3c23d015-ef00-4a30-b2da-dc15d140a773","Type":"ContainerStarted","Data":"eb9f4d20351befc0a1a9aaa6bdd3d2aff320a7818cedbfc9c7141aeeb2ceb8bb"} Oct 11 03:53:35 crc kubenswrapper[4967]: I1011 03:53:35.990303 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rv6n2" event={"ID":"c3a20ddb-cfaa-46ac-8aed-4db9de3e32b6","Type":"ContainerStarted","Data":"ba6eda2a6f7f478ffc22cba3ee82f6a21f7e1046b4e827daf8e521630821e2b6"} Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.015223 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-7sxxj" event={"ID":"45ebd02f-dc19-48a5-8309-bdbf11a8a7a4","Type":"ContainerStarted","Data":"fa81e48ec760f4197f941136b6a03f72f05ce5908ba54448de5b84d7defac8b4"} Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.048613 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tr7z7" event={"ID":"092da80d-ee9c-4e88-b3e2-58bb3dd77f9b","Type":"ContainerStarted","Data":"33dfeaff691326d3f3bc1735d64777d4f82f706fd7df5b3d8d381c12d561d6ea"} Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.048658 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tr7z7" event={"ID":"092da80d-ee9c-4e88-b3e2-58bb3dd77f9b","Type":"ContainerStarted","Data":"8b84d45c585d6f8c52d51403a21febb8fef40eb2a08792f5662f04c22a7fb31b"} Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.064474 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:36 crc kubenswrapper[4967]: E1011 03:53:36.071438 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:36.571420406 +0000 UTC m=+144.534629339 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.071545 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:36 crc kubenswrapper[4967]: E1011 03:53:36.071895 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:36.571887698 +0000 UTC m=+144.535096631 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.084694 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n6qrf" event={"ID":"7fa0bf77-28c4-492d-a4d4-b0a0ec401cf0","Type":"ContainerStarted","Data":"150326476e627d3435f1faa00bf03497b879829b677646b6cb2a499318a5b1d5"} Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.088006 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gbqfq" event={"ID":"81c6feb2-e6ea-43f0-834d-499bea65b4aa","Type":"ContainerStarted","Data":"cd4a3f5c52710044f73dc56a069400f989f561110a86dd2410c9429b9aef51c4"} Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.114122 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-csl6r" event={"ID":"5e20a512-8f34-4d21-be45-e749f7a6cc87","Type":"ContainerStarted","Data":"e8cf8cba0e027047127e72821d4240c617453de6ea5d373ed64d907303c53a6e"} Oct 11 03:53:36 crc kubenswrapper[4967]: W1011 03:53:36.123323 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode1a3f43f_63fe_4816_8414_2d7e4c63be57.slice/crio-a744031609705804e5e6789db23e14d43e3737af69a491a039850d7a71492d19 WatchSource:0}: Error finding container a744031609705804e5e6789db23e14d43e3737af69a491a039850d7a71492d19: Status 404 returned error can't find the container with id a744031609705804e5e6789db23e14d43e3737af69a491a039850d7a71492d19 Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.126648 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cwz9p" event={"ID":"038209e7-3fdb-490b-976e-f1791bff29d8","Type":"ContainerStarted","Data":"47733e3d9c4a61e886a58265c1811cb6c0dee91c9762d513c6d6aa2e18758d79"} Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.160306 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ws6r9" event={"ID":"d89bf2d0-8109-40be-b1ce-26ee92a58d81","Type":"ContainerStarted","Data":"6c40743bce639dc44a931dc104e22cf561b884680faf9f718d1d11dc83ac91fb"} Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.160366 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ws6r9" event={"ID":"d89bf2d0-8109-40be-b1ce-26ee92a58d81","Type":"ContainerStarted","Data":"513531be0dbf16f91c5424ecad9c48628d2e4f8525fd5e5cbec669843390eb23"} Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.169441 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kmpsc" event={"ID":"49347426-bca5-4fde-8b87-21dc8fc486d8","Type":"ContainerStarted","Data":"ce7690ef3ffe0ef8173cc01455015a8410084c644d91cc2315c8aa27ffacb2db"} Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.177704 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:36 crc kubenswrapper[4967]: E1011 03:53:36.178900 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:36.678881299 +0000 UTC m=+144.642090232 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.203776 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rmkqj" event={"ID":"3883197f-52dd-465a-82bb-c4dd32b3d090","Type":"ContainerStarted","Data":"7f991df4008ad12895a66dd07792a4295c2ca53b880205e297c528a4eaf530f5"} Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.278341 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-49rxs" event={"ID":"b9509019-37bf-4674-af71-c259393eab0b","Type":"ContainerStarted","Data":"09dd985e0fa0ae2f2cbb5ea8523da21f78bc2436ad959c2f3bf1d8b812b2e01d"} Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.278908 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:36 crc kubenswrapper[4967]: E1011 03:53:36.279433 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:36.779418463 +0000 UTC m=+144.742627396 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.279991 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gwrxz" event={"ID":"ac8fc0d4-aa55-4974-a506-7ccd2713f31a","Type":"ContainerStarted","Data":"b9691e0411e7c08a664e730f8a35b0871c778537e683559cd093dfd9ac05ba24"} Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.314598 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-rmkqj" podStartSLOduration=123.314585703 podStartE2EDuration="2m3.314585703s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:36.312594299 +0000 UTC m=+144.275803232" watchObservedRunningTime="2025-10-11 03:53:36.314585703 +0000 UTC m=+144.277794636" Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.339514 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-dtj88" event={"ID":"d6b9a450-fd86-47a3-9c0e-f61aa7a24377","Type":"ContainerStarted","Data":"9e23760ef5636af900f586efabf514fdb67a1b319c96e4e989f818533ca865cc"} Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.339869 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-dtj88" Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.343013 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" event={"ID":"0dae89fc-445f-4531-8928-29b334636522","Type":"ContainerStarted","Data":"21a94896d91097d19aeb6b46354b56d5640a05428b63933176e0a1178bb5f651"} Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.343860 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.345334 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-kdpld" event={"ID":"c506a84b-4ea6-477e-88de-fceec3a95f31","Type":"ContainerStarted","Data":"4c084fa9b84e40f8ef355f65935253cecd1d31951d86121a4cfbd6bc4b98453a"} Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.349879 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cp2n5" event={"ID":"7b200d8c-28e7-494b-99ec-893f1d9ba3ca","Type":"ContainerStarted","Data":"dbd44a03ee04bd8a6279209a5de19c85dffb8d4753bb4e763cb11fc60d213e6d"} Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.352738 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bjc5w" event={"ID":"8f4cc87b-f2fd-4c61-82fb-f362b81445d7","Type":"ContainerStarted","Data":"d1776011cb5c7c035cd896689d91805cc226ac57e8ca28caa0bf7c7aef2dc753"} Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.360508 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-kmpsc" podStartSLOduration=123.360493167 podStartE2EDuration="2m3.360493167s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:36.360310982 +0000 UTC m=+144.323519915" watchObservedRunningTime="2025-10-11 03:53:36.360493167 +0000 UTC m=+144.323702100" Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.364234 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-9jjwn" event={"ID":"b5fca2ad-c70a-43e2-aff1-f180617a94f8","Type":"ContainerStarted","Data":"2338fd808e682c034f980d128b44242e71866aae1cea7eb20e11718bcaa722b5"} Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.367531 4967 patch_prober.go:28] interesting pod/downloads-7954f5f757-dtj88 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.367568 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-dtj88" podUID="d6b9a450-fd86-47a3-9c0e-f61aa7a24377" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.376521 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-fsc8x" event={"ID":"23902dde-0c8d-4d54-99f4-0ee4e19a1a50","Type":"ContainerStarted","Data":"8eec49ef7cb5a70080f7c27281e3f79b54801bcd4416ebdfdf991e5a070de733"} Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.379473 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-g4q8m" event={"ID":"47a99167-adc4-40b4-9925-778c670ad55f","Type":"ContainerStarted","Data":"0c126d4600570a5a9e3ee078cd37f79a2f3caa2175119913e08b476484f98682"} Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.393560 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:36 crc kubenswrapper[4967]: E1011 03:53:36.393847 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:36.893831167 +0000 UTC m=+144.857040100 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.393890 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.394733 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-2fvpq" event={"ID":"0761866e-d6ab-40c9-a776-8c7fb2683fad","Type":"ContainerStarted","Data":"2c152b9bb6aaf6ca21e92303d82c058b46ee968ec18647be2a62c2114f5b9103"} Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.394779 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-2fvpq" event={"ID":"0761866e-d6ab-40c9-a776-8c7fb2683fad","Type":"ContainerStarted","Data":"3a1a4bbfa346fdb02307a46cef0ee3f6ff8eb1b049cfa6b21075a19211cbca38"} Oct 11 03:53:36 crc kubenswrapper[4967]: E1011 03:53:36.395182 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:36.895173983 +0000 UTC m=+144.858382916 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.410589 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mhnj2" event={"ID":"bc099334-ba8a-4538-9a6d-98084ecb9a63","Type":"ContainerStarted","Data":"42b5ae4642ccf8c4dbfe4d2cd557fabc6bb11e7b76d259c9c13fc81a3ae4dc67"} Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.417725 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-9c49l" event={"ID":"315ce8c5-1d03-4b22-885d-04d07e83ab9d","Type":"ContainerStarted","Data":"8e262c1cae687f8e390984d583a76611661fab821d81fbb6a36a171240d19b3c"} Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.430481 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-7sxxj" podStartSLOduration=123.430440046 podStartE2EDuration="2m3.430440046s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:36.427778983 +0000 UTC m=+144.390987916" watchObservedRunningTime="2025-10-11 03:53:36.430440046 +0000 UTC m=+144.393648979" Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.431430 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-92sz4" Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.495020 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:36 crc kubenswrapper[4967]: E1011 03:53:36.497152 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:36.997136227 +0000 UTC m=+144.960345160 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.515099 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnjpj"] Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.543303 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-csl6r" podStartSLOduration=123.543287487 podStartE2EDuration="2m3.543287487s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:36.529447089 +0000 UTC m=+144.492656022" watchObservedRunningTime="2025-10-11 03:53:36.543287487 +0000 UTC m=+144.506496410" Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.543884 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v7xvv"] Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.566140 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-jg6t8"] Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.566825 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-rt5vd" Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.567043 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-mrsjk" podStartSLOduration=123.567033545 podStartE2EDuration="2m3.567033545s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:36.566908692 +0000 UTC m=+144.530117625" watchObservedRunningTime="2025-10-11 03:53:36.567033545 +0000 UTC m=+144.530242478" Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.579253 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-4xkx2"] Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.597592 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:36 crc kubenswrapper[4967]: E1011 03:53:36.597867 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:37.097856947 +0000 UTC m=+145.061065880 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.615105 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rv6n2" podStartSLOduration=123.615090877 podStartE2EDuration="2m3.615090877s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:36.613406581 +0000 UTC m=+144.576615514" watchObservedRunningTime="2025-10-11 03:53:36.615090877 +0000 UTC m=+144.578299810" Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.644216 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-tr7z7" podStartSLOduration=123.644197671 podStartE2EDuration="2m3.644197671s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:36.643643567 +0000 UTC m=+144.606852500" watchObservedRunningTime="2025-10-11 03:53:36.644197671 +0000 UTC m=+144.607406604" Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.682650 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-dtj88" podStartSLOduration=123.68263487 podStartE2EDuration="2m3.68263487s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:36.68039569 +0000 UTC m=+144.643604623" watchObservedRunningTime="2025-10-11 03:53:36.68263487 +0000 UTC m=+144.645843793" Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.705918 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:36 crc kubenswrapper[4967]: E1011 03:53:36.706012 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:37.205997548 +0000 UTC m=+145.169206481 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.706190 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:36 crc kubenswrapper[4967]: E1011 03:53:36.706491 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:37.206473981 +0000 UTC m=+145.169682914 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.715398 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-2fvpq" podStartSLOduration=123.715380955 podStartE2EDuration="2m3.715380955s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:36.714517491 +0000 UTC m=+144.677726424" watchObservedRunningTime="2025-10-11 03:53:36.715380955 +0000 UTC m=+144.678589888" Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.748801 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-bjc5w" podStartSLOduration=123.748784807 podStartE2EDuration="2m3.748784807s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:36.7474341 +0000 UTC m=+144.710643043" watchObservedRunningTime="2025-10-11 03:53:36.748784807 +0000 UTC m=+144.711993740" Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.807260 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:36 crc kubenswrapper[4967]: E1011 03:53:36.807821 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:37.307800868 +0000 UTC m=+145.271009801 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.827895 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gwrxz" podStartSLOduration=123.827873305 podStartE2EDuration="2m3.827873305s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:36.819956059 +0000 UTC m=+144.783164982" watchObservedRunningTime="2025-10-11 03:53:36.827873305 +0000 UTC m=+144.791082238" Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.892006 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-9c49l" podStartSLOduration=123.891985746 podStartE2EDuration="2m3.891985746s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:36.879626978 +0000 UTC m=+144.842835911" watchObservedRunningTime="2025-10-11 03:53:36.891985746 +0000 UTC m=+144.855194679" Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.918986 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:36 crc kubenswrapper[4967]: E1011 03:53:36.919358 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:37.419346392 +0000 UTC m=+145.382555325 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:36 crc kubenswrapper[4967]: I1011 03:53:36.923000 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" podStartSLOduration=123.922977411 podStartE2EDuration="2m3.922977411s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:36.913297957 +0000 UTC m=+144.876506890" watchObservedRunningTime="2025-10-11 03:53:36.922977411 +0000 UTC m=+144.886186344" Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.020000 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:37 crc kubenswrapper[4967]: E1011 03:53:37.020322 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:37.520304559 +0000 UTC m=+145.483513492 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.121984 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:37 crc kubenswrapper[4967]: E1011 03:53:37.123950 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:37.623922687 +0000 UTC m=+145.587131620 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.222670 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:37 crc kubenswrapper[4967]: E1011 03:53:37.222963 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:37.72294926 +0000 UTC m=+145.686158193 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.256978 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-2fvpq" Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.271080 4967 patch_prober.go:28] interesting pod/router-default-5444994796-2fvpq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 11 03:53:37 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Oct 11 03:53:37 crc kubenswrapper[4967]: [+]process-running ok Oct 11 03:53:37 crc kubenswrapper[4967]: healthz check failed Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.271121 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2fvpq" podUID="0761866e-d6ab-40c9-a776-8c7fb2683fad" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.326899 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:37 crc kubenswrapper[4967]: E1011 03:53:37.327218 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:37.827207307 +0000 UTC m=+145.790416230 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.344416 4967 patch_prober.go:28] interesting pod/oauth-openshift-558db77b4-6sjfd container/oauth-openshift namespace/openshift-authentication: Readiness probe status=failure output="Get \"https://10.217.0.15:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.347592 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" podUID="0dae89fc-445f-4531-8928-29b334636522" containerName="oauth-openshift" probeResult="failure" output="Get \"https://10.217.0.15:6443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.432409 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:37 crc kubenswrapper[4967]: E1011 03:53:37.432797 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:37.932781728 +0000 UTC m=+145.895990661 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.508336 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bxjls" event={"ID":"89f0e615-1fde-483d-9b38-83287f104360","Type":"ContainerStarted","Data":"493f51d096d6d3377321b89068bad1f631307d3c6bef405cc111ec4589936111"} Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.509562 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bxjls" Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.522417 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-jg6t8" event={"ID":"1b95aa73-24a0-496b-988c-5f0d52f1f00e","Type":"ContainerStarted","Data":"5d241dcab4c648c22d5c3854a12abec6b78b37f88b7870d71ea79f7d4cb173cf"} Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.535896 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:37 crc kubenswrapper[4967]: E1011 03:53:37.536228 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:38.036216882 +0000 UTC m=+145.999425805 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.543294 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rw8ng" event={"ID":"b173cdfc-6a31-4d2e-93cd-a04eb1c68877","Type":"ContainerStarted","Data":"f6cdc210b695837d3a09c039bfb80596a7eeda22434e5ebd0696ec2e8c20d013"} Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.566799 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bxjls" Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.568498 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-bxjls" podStartSLOduration=124.568477463 podStartE2EDuration="2m4.568477463s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:37.566612472 +0000 UTC m=+145.529821405" watchObservedRunningTime="2025-10-11 03:53:37.568477463 +0000 UTC m=+145.531686396" Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.581223 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnjpj" event={"ID":"e6b311ab-c933-4950-9104-8071d75d77ac","Type":"ContainerStarted","Data":"4be42c6e81de39682f36e833ee9482a16406a7f3bedc5fe85e6771272b1defd2"} Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.581264 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnjpj" event={"ID":"e6b311ab-c933-4950-9104-8071d75d77ac","Type":"ContainerStarted","Data":"b6a7328b86ab051220902573ce39c07457d1269a4e1b509a30e32b40883865c9"} Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.581612 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnjpj" Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.589918 4967 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-wnjpj container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.22:5443/healthz\": dial tcp 10.217.0.22:5443: connect: connection refused" start-of-body= Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.589964 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnjpj" podUID="e6b311ab-c933-4950-9104-8071d75d77ac" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.22:5443/healthz\": dial tcp 10.217.0.22:5443: connect: connection refused" Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.604377 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cp2n5" event={"ID":"7b200d8c-28e7-494b-99ec-893f1d9ba3ca","Type":"ContainerStarted","Data":"a57b7503b29a37866369c867168e5116d305566cd24c4c0f878be03291205777"} Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.604416 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cp2n5" event={"ID":"7b200d8c-28e7-494b-99ec-893f1d9ba3ca","Type":"ContainerStarted","Data":"766f46415f99ba4cb37fd6c555ff90fcd2549dfb0999c6e710f8435a2804b700"} Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.628244 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-dnnwg" event={"ID":"e1a3f43f-63fe-4816-8414-2d7e4c63be57","Type":"ContainerStarted","Data":"a744031609705804e5e6789db23e14d43e3737af69a491a039850d7a71492d19"} Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.637583 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:37 crc kubenswrapper[4967]: E1011 03:53:37.638631 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:38.138616677 +0000 UTC m=+146.101825610 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.701142 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" event={"ID":"454804e7-0d4d-4536-8544-87a5b6368f70","Type":"ContainerStarted","Data":"da377bf0ad7c79df5c37821868424d436f221df9c652d59b3cab9c1c94579517"} Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.736993 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnjpj" podStartSLOduration=124.736978512 podStartE2EDuration="2m4.736978512s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:37.736577651 +0000 UTC m=+145.699786584" watchObservedRunningTime="2025-10-11 03:53:37.736978512 +0000 UTC m=+145.700187445" Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.738545 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-rw8ng" podStartSLOduration=124.738522204 podStartE2EDuration="2m4.738522204s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:37.685573079 +0000 UTC m=+145.648782012" watchObservedRunningTime="2025-10-11 03:53:37.738522204 +0000 UTC m=+145.701731137" Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.746354 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.746566 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n6qrf" event={"ID":"7fa0bf77-28c4-492d-a4d4-b0a0ec401cf0","Type":"ContainerStarted","Data":"f391e039f8df526bba352a8de2fc9e269b6607f00e2622cda0cf98d16e25ee10"} Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.746662 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n6qrf" event={"ID":"7fa0bf77-28c4-492d-a4d4-b0a0ec401cf0","Type":"ContainerStarted","Data":"4184948b9d9db56383f1ef0b4fde4dca26f67ab2755a06e570639639c92e0061"} Oct 11 03:53:37 crc kubenswrapper[4967]: E1011 03:53:37.747370 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:38.247356386 +0000 UTC m=+146.210565319 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.772845 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-dtj88" event={"ID":"d6b9a450-fd86-47a3-9c0e-f61aa7a24377","Type":"ContainerStarted","Data":"c81ce703200487e68a9614d9f741f57a5f84522d9b06115f7b9d97984653e302"} Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.782998 4967 patch_prober.go:28] interesting pod/downloads-7954f5f757-dtj88 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.783036 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-dtj88" podUID="d6b9a450-fd86-47a3-9c0e-f61aa7a24377" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.821566 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wtn79" event={"ID":"5449c948-487e-426d-b151-31e8079576f3","Type":"ContainerStarted","Data":"71fb1ca53ba385ae5087477e3e0fee48dbe734d830c5e65433013adbbd6f396c"} Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.822355 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wtn79" Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.825035 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-cp2n5" podStartSLOduration=124.825019715 podStartE2EDuration="2m4.825019715s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:37.822489276 +0000 UTC m=+145.785698209" watchObservedRunningTime="2025-10-11 03:53:37.825019715 +0000 UTC m=+145.788228648" Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.842395 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-fsc8x" event={"ID":"23902dde-0c8d-4d54-99f4-0ee4e19a1a50","Type":"ContainerStarted","Data":"69e8b532f98acde882d7bb219ff805f7d1ae0deb2e4d5dbf0db55e4b4f3fb6ed"} Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.850349 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:37 crc kubenswrapper[4967]: E1011 03:53:37.851157 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:38.351140669 +0000 UTC m=+146.314349602 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.864437 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" event={"ID":"efe9c94d-8f97-4274-a494-a5ecf5b88cec","Type":"ContainerStarted","Data":"249854c9157c05ef244b509d9d49e3eca7dd2d2d9e145e7a2c5f4b457b2ad5f7"} Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.890599 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-n6qrf" podStartSLOduration=124.890579295 podStartE2EDuration="2m4.890579295s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:37.867508696 +0000 UTC m=+145.830717629" watchObservedRunningTime="2025-10-11 03:53:37.890579295 +0000 UTC m=+145.853788228" Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.892401 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wtn79" podStartSLOduration=124.892394954 podStartE2EDuration="2m4.892394954s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:37.892294052 +0000 UTC m=+145.855502985" watchObservedRunningTime="2025-10-11 03:53:37.892394954 +0000 UTC m=+145.855603887" Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.924554 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gbqfq" event={"ID":"81c6feb2-e6ea-43f0-834d-499bea65b4aa","Type":"ContainerStarted","Data":"6f057e4f5d06d8ddbe7f474a6dcc9927f5037e425ae99deba44c680cd47fa07a"} Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.943235 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-rv6n2" event={"ID":"c3a20ddb-cfaa-46ac-8aed-4db9de3e32b6","Type":"ContainerStarted","Data":"f1e96511c457469fbd0e41baa85604c2bd243fd4b81a2bd08c6920cb5e2d76c4"} Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.951576 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:37 crc kubenswrapper[4967]: E1011 03:53:37.953278 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:38.453265967 +0000 UTC m=+146.416474900 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.968837 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v7xvv" event={"ID":"ee299c82-155e-4da8-9d33-80c85d5460d9","Type":"ContainerStarted","Data":"e4e888bdae47ad7c5891f6872b44cad7e3d99230fa4afdeec25d1ad16ed96fa6"} Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.969513 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v7xvv" Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.987039 4967 patch_prober.go:28] interesting pod/catalog-operator-68c6474976-v7xvv container/catalog-operator namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" start-of-body= Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.987204 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v7xvv" podUID="ee299c82-155e-4da8-9d33-80c85d5460d9" containerName="catalog-operator" probeResult="failure" output="Get \"https://10.217.0.29:8443/healthz\": dial tcp 10.217.0.29:8443: connect: connection refused" Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.995754 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" podStartSLOduration=124.995741536 podStartE2EDuration="2m4.995741536s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:37.954057728 +0000 UTC m=+145.917266661" watchObservedRunningTime="2025-10-11 03:53:37.995741536 +0000 UTC m=+145.958950469" Oct 11 03:53:37 crc kubenswrapper[4967]: I1011 03:53:37.996581 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gbqfq" podStartSLOduration=124.996574459 podStartE2EDuration="2m4.996574459s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:37.995453728 +0000 UTC m=+145.958662661" watchObservedRunningTime="2025-10-11 03:53:37.996574459 +0000 UTC m=+145.959783382" Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.002246 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mhnj2" event={"ID":"bc099334-ba8a-4538-9a6d-98084ecb9a63","Type":"ContainerStarted","Data":"61eb14b2c794236c58e4ad7438edb9aacdaa52d1aed2c4cb883e940baa34fee0"} Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.002712 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mhnj2" Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.055847 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:38 crc kubenswrapper[4967]: E1011 03:53:38.056852 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:38.556835584 +0000 UTC m=+146.520044517 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.058900 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-9jjwn" event={"ID":"b5fca2ad-c70a-43e2-aff1-f180617a94f8","Type":"ContainerStarted","Data":"030567f704a878ab88d96a52fdcdb434bb2e14ebd6628b1efd23c5a14bc56fd3"} Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.072730 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-d8x8n" event={"ID":"3c23d015-ef00-4a30-b2da-dc15d140a773","Type":"ContainerStarted","Data":"9c595bb84fa54f4fcf3b5de5f34cd86945ef329517ad492a9c5676a542a7b244"} Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.079242 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-4xkx2" event={"ID":"de60a4e8-2c49-4b5b-a28d-b782c0852911","Type":"ContainerStarted","Data":"44c4c0fbb09a02e955f22d4b00fa61582a84e7a6bda1aeaff3d6a0c37db74f63"} Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.083841 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mhnj2" podStartSLOduration=125.083829111 podStartE2EDuration="2m5.083829111s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:38.083490102 +0000 UTC m=+146.046699035" watchObservedRunningTime="2025-10-11 03:53:38.083829111 +0000 UTC m=+146.047038034" Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.084675 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-fsc8x" podStartSLOduration=7.084669914 podStartE2EDuration="7.084669914s" podCreationTimestamp="2025-10-11 03:53:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:38.026502666 +0000 UTC m=+145.989711589" watchObservedRunningTime="2025-10-11 03:53:38.084669914 +0000 UTC m=+146.047878847" Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.092352 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7nk8v" event={"ID":"22ab7539-f66a-421e-8053-1cd8e03d465f","Type":"ContainerStarted","Data":"81199afcff176912a59914d74b8a3c86f0d54aee379c2222213362dd1181aa27"} Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.103761 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wtn79" Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.117260 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-9c49l" event={"ID":"315ce8c5-1d03-4b22-885d-04d07e83ab9d","Type":"ContainerStarted","Data":"30bdbe3ba302bc9e46d97c42ce9fee7f7cbacfc19a6723800f279b1242afcee6"} Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.128368 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v7xvv" podStartSLOduration=125.128353996 podStartE2EDuration="2m5.128353996s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:38.126471255 +0000 UTC m=+146.089680188" watchObservedRunningTime="2025-10-11 03:53:38.128353996 +0000 UTC m=+146.091562929" Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.144479 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-9jjwn" podStartSLOduration=125.144465165 podStartE2EDuration="2m5.144465165s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:38.14318175 +0000 UTC m=+146.106390683" watchObservedRunningTime="2025-10-11 03:53:38.144465165 +0000 UTC m=+146.107674098" Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.151435 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-g4q8m" event={"ID":"47a99167-adc4-40b4-9925-778c670ad55f","Type":"ContainerStarted","Data":"ddc8cad3077b4b0b9050e74e60730e2d87998c3895d157489a7977c6b2f2f5c9"} Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.157162 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:38 crc kubenswrapper[4967]: E1011 03:53:38.158844 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:38.658832819 +0000 UTC m=+146.622041752 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.175839 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-kdpld" event={"ID":"c506a84b-4ea6-477e-88de-fceec3a95f31","Type":"ContainerStarted","Data":"cc575196dfaaf2394c082a5d499f0ffaae1a90e9adfdc0626c04af2ad9f0fcba"} Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.199457 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cwz9p" event={"ID":"038209e7-3fdb-490b-976e-f1791bff29d8","Type":"ContainerStarted","Data":"7113d777d5c70e005ac19cfbea1f8dabd6c163ee1747592deb08b55a77b0bb37"} Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.207782 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-4xkx2" podStartSLOduration=125.207767074 podStartE2EDuration="2m5.207767074s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:38.180489819 +0000 UTC m=+146.143698762" watchObservedRunningTime="2025-10-11 03:53:38.207767074 +0000 UTC m=+146.170976007" Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.238957 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ws79r" event={"ID":"14860b75-6fa1-43b5-ba20-d731531dba3b","Type":"ContainerStarted","Data":"d38f3bfd2fbea11c1abe1ad260d7a5afad0604b2c5c21fa1415cd58ae2b91927"} Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.239005 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ws79r" event={"ID":"14860b75-6fa1-43b5-ba20-d731531dba3b","Type":"ContainerStarted","Data":"cb0a8e0abfdf5878be4120090a5d926a0a96399b35595638f044f1d3277c3b24"} Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.247135 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7nk8v" podStartSLOduration=125.247123019 podStartE2EDuration="2m5.247123019s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:38.246940284 +0000 UTC m=+146.210149207" watchObservedRunningTime="2025-10-11 03:53:38.247123019 +0000 UTC m=+146.210331952" Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.253672 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-49rxs" event={"ID":"b9509019-37bf-4674-af71-c259393eab0b","Type":"ContainerStarted","Data":"0537240f76748e255545f491ddc4f3f03411c1717cb8e8b271954b3d53028eff"} Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.254660 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-49rxs" Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.258679 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:38 crc kubenswrapper[4967]: E1011 03:53:38.259405 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:38.759383762 +0000 UTC m=+146.722592695 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.265151 4967 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-49rxs container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/healthz\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.265538 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-49rxs" podUID="b9509019-37bf-4674-af71-c259393eab0b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.25:8080/healthz\": dial tcp 10.217.0.25:8080: connect: connection refused" Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.265930 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwp4c" event={"ID":"943c9c2f-24f5-408c-8e8d-c344253fc8f6","Type":"ContainerStarted","Data":"224c5b7012a6b5bea59c2c124c05d2a39678782088607104578cef0b372567f6"} Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.266657 4967 patch_prober.go:28] interesting pod/router-default-5444994796-2fvpq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 11 03:53:38 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Oct 11 03:53:38 crc kubenswrapper[4967]: [+]process-running ok Oct 11 03:53:38 crc kubenswrapper[4967]: healthz check failed Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.267038 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2fvpq" podUID="0761866e-d6ab-40c9-a776-8c7fb2683fad" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.290549 4967 generic.go:334] "Generic (PLEG): container finished" podID="d89bf2d0-8109-40be-b1ce-26ee92a58d81" containerID="6c40743bce639dc44a931dc104e22cf561b884680faf9f718d1d11dc83ac91fb" exitCode=0 Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.291514 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ws6r9" event={"ID":"d89bf2d0-8109-40be-b1ce-26ee92a58d81","Type":"ContainerDied","Data":"6c40743bce639dc44a931dc104e22cf561b884680faf9f718d1d11dc83ac91fb"} Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.291733 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ws6r9" Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.291823 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ws6r9" event={"ID":"d89bf2d0-8109-40be-b1ce-26ee92a58d81","Type":"ContainerStarted","Data":"edf38c9ad987dd3c13e2cc3018e736e638d657fd2e3cc11ff2d25aa286841189"} Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.292600 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.295830 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.319681 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.350460 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-kdpld" podStartSLOduration=125.350440749 podStartE2EDuration="2m5.350440749s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:38.346963864 +0000 UTC m=+146.310172797" watchObservedRunningTime="2025-10-11 03:53:38.350440749 +0000 UTC m=+146.313649682" Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.351483 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-cwz9p" podStartSLOduration=125.351477337 podStartE2EDuration="2m5.351477337s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:38.274608118 +0000 UTC m=+146.237817051" watchObservedRunningTime="2025-10-11 03:53:38.351477337 +0000 UTC m=+146.314686270" Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.361363 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:38 crc kubenswrapper[4967]: E1011 03:53:38.369090 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:38.865091779 +0000 UTC m=+146.828300712 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.380685 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ws79r" podStartSLOduration=125.380671154 podStartE2EDuration="2m5.380671154s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:38.374063803 +0000 UTC m=+146.337272736" watchObservedRunningTime="2025-10-11 03:53:38.380671154 +0000 UTC m=+146.343880077" Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.462543 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:38 crc kubenswrapper[4967]: E1011 03:53:38.462862 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:38.962847137 +0000 UTC m=+146.926056060 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.534375 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ws6r9" podStartSLOduration=125.534358179 podStartE2EDuration="2m5.534358179s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:38.495878658 +0000 UTC m=+146.459087591" watchObservedRunningTime="2025-10-11 03:53:38.534358179 +0000 UTC m=+146.497567102" Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.551786 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-vwp4c" podStartSLOduration=125.551768974 podStartE2EDuration="2m5.551768974s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:38.551476637 +0000 UTC m=+146.514685570" watchObservedRunningTime="2025-10-11 03:53:38.551768974 +0000 UTC m=+146.514977897" Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.553394 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-49rxs" podStartSLOduration=125.553385948 podStartE2EDuration="2m5.553385948s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:38.534341249 +0000 UTC m=+146.497550182" watchObservedRunningTime="2025-10-11 03:53:38.553385948 +0000 UTC m=+146.516594881" Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.565048 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:38 crc kubenswrapper[4967]: E1011 03:53:38.565386 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:39.065372226 +0000 UTC m=+147.028581159 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.567258 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.665964 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:38 crc kubenswrapper[4967]: E1011 03:53:38.666315 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:39.166299571 +0000 UTC m=+147.129508504 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.767245 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:38 crc kubenswrapper[4967]: E1011 03:53:38.767666 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:39.267648507 +0000 UTC m=+147.230857440 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.868229 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:38 crc kubenswrapper[4967]: E1011 03:53:38.868431 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:39.368404308 +0000 UTC m=+147.331613241 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.868558 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:38 crc kubenswrapper[4967]: E1011 03:53:38.868849 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:39.36883625 +0000 UTC m=+147.332045183 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.969857 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:38 crc kubenswrapper[4967]: E1011 03:53:38.969987 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:39.46996122 +0000 UTC m=+147.433170153 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:38 crc kubenswrapper[4967]: I1011 03:53:38.970148 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:38 crc kubenswrapper[4967]: E1011 03:53:38.970481 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:39.470474054 +0000 UTC m=+147.433682987 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.070955 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:39 crc kubenswrapper[4967]: E1011 03:53:39.071109 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:39.571082051 +0000 UTC m=+147.534291004 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.071145 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:39 crc kubenswrapper[4967]: E1011 03:53:39.071467 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:39.571459061 +0000 UTC m=+147.534667994 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.172196 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:39 crc kubenswrapper[4967]: E1011 03:53:39.172449 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:39.672425308 +0000 UTC m=+147.635634241 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.172758 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:39 crc kubenswrapper[4967]: E1011 03:53:39.173215 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:39.673201368 +0000 UTC m=+147.636410301 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.233737 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.265298 4967 patch_prober.go:28] interesting pod/router-default-5444994796-2fvpq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 11 03:53:39 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Oct 11 03:53:39 crc kubenswrapper[4967]: [+]process-running ok Oct 11 03:53:39 crc kubenswrapper[4967]: healthz check failed Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.265358 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2fvpq" podUID="0761866e-d6ab-40c9-a776-8c7fb2683fad" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.273357 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:39 crc kubenswrapper[4967]: E1011 03:53:39.273530 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:39.773508247 +0000 UTC m=+147.736717180 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.273656 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:39 crc kubenswrapper[4967]: E1011 03:53:39.273979 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:39.773970869 +0000 UTC m=+147.737179802 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.296757 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mhnj2" event={"ID":"bc099334-ba8a-4538-9a6d-98084ecb9a63","Type":"ContainerStarted","Data":"5c0c8d0306ce4aac14f5403837866c1b30a58e821c8735441b64b523ca44c1df"} Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.298538 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-dnnwg" event={"ID":"e1a3f43f-63fe-4816-8414-2d7e4c63be57","Type":"ContainerStarted","Data":"874f4fe667ee820b8c07c8800c52eeaa77db6051c7ca5a50f455a089b4f6d973"} Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.299827 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-d8x8n" event={"ID":"3c23d015-ef00-4a30-b2da-dc15d140a773","Type":"ContainerStarted","Data":"fc9393a70163f9ba421fc382e0af5aeff0a776cbee1badf04e79e5c37273882c"} Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.300509 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-d8x8n" Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.301901 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-ws79r" event={"ID":"14860b75-6fa1-43b5-ba20-d731531dba3b","Type":"ContainerStarted","Data":"373960a74a6b2dd4f63d8c877a4523ec081bb2557f7d7053c453c5715006f696"} Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.304115 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-4xkx2" event={"ID":"de60a4e8-2c49-4b5b-a28d-b782c0852911","Type":"ContainerStarted","Data":"498f7988317682b1728473f0adc91b6344f8c3792518371e3319957a0ed0428e"} Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.305701 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-jg6t8" event={"ID":"1b95aa73-24a0-496b-988c-5f0d52f1f00e","Type":"ContainerStarted","Data":"85ccfb73b93e411c0b3f478c562be5b3b4d11a085ae59b96b1f5083bf32af4c8"} Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.307321 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-g4q8m" event={"ID":"47a99167-adc4-40b4-9925-778c670ad55f","Type":"ContainerStarted","Data":"645f41a31ca4c2320248a68afa40c86bc0fdcad072c27b6dbb8b73c8bd4845d0"} Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.309389 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v7xvv" event={"ID":"ee299c82-155e-4da8-9d33-80c85d5460d9","Type":"ContainerStarted","Data":"772d1d73adfecb2e2789e8e5dde087d21a01034e6cc596503b69d062f5e29c76"} Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.318858 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" event={"ID":"efe9c94d-8f97-4274-a494-a5ecf5b88cec","Type":"ContainerStarted","Data":"f9d2a9282e0b232fdced156bee092ae53b4fa2f53af6b1869ad721fbde30c946"} Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.322128 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-gbqfq" event={"ID":"81c6feb2-e6ea-43f0-834d-499bea65b4aa","Type":"ContainerStarted","Data":"12e23a5cc9bb496ed05ab81ab5d23c65a394f3e6a09ba340dc36831697de1a49"} Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.322178 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-v7xvv" Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.324371 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-7nk8v" event={"ID":"22ab7539-f66a-421e-8053-1cd8e03d465f","Type":"ContainerStarted","Data":"93f3edbaefb84d4389965315e7665a26b679d5563bb840118522bf13c8938843"} Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.325295 4967 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-49rxs container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.25:8080/healthz\": dial tcp 10.217.0.25:8080: connect: connection refused" start-of-body= Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.325326 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-49rxs" podUID="b9509019-37bf-4674-af71-c259393eab0b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.25:8080/healthz\": dial tcp 10.217.0.25:8080: connect: connection refused" Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.326413 4967 patch_prober.go:28] interesting pod/downloads-7954f5f757-dtj88 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.326436 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-dtj88" podUID="d6b9a450-fd86-47a3-9c0e-f61aa7a24377" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.336570 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-nmzrx" Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.344852 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-g4q8m" podStartSLOduration=126.344835854 podStartE2EDuration="2m6.344835854s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:39.343138578 +0000 UTC m=+147.306347511" watchObservedRunningTime="2025-10-11 03:53:39.344835854 +0000 UTC m=+147.308044787" Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.345906 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-d8x8n" podStartSLOduration=8.345900462 podStartE2EDuration="8.345900462s" podCreationTimestamp="2025-10-11 03:53:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:39.316318095 +0000 UTC m=+147.279527028" watchObservedRunningTime="2025-10-11 03:53:39.345900462 +0000 UTC m=+147.309109395" Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.375757 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:39 crc kubenswrapper[4967]: E1011 03:53:39.377796 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:39.877770863 +0000 UTC m=+147.840979796 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.395693 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" podStartSLOduration=126.395675132 podStartE2EDuration="2m6.395675132s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:39.393179143 +0000 UTC m=+147.356388076" watchObservedRunningTime="2025-10-11 03:53:39.395675132 +0000 UTC m=+147.358884065" Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.397693 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-jg6t8" podStartSLOduration=8.397684386 podStartE2EDuration="8.397684386s" podCreationTimestamp="2025-10-11 03:53:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:39.366260039 +0000 UTC m=+147.329468972" watchObservedRunningTime="2025-10-11 03:53:39.397684386 +0000 UTC m=+147.360893319" Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.484759 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:39 crc kubenswrapper[4967]: E1011 03:53:39.485098 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:39.985087023 +0000 UTC m=+147.948295956 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.586821 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:39 crc kubenswrapper[4967]: E1011 03:53:39.587627 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:40.087609441 +0000 UTC m=+148.050818374 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.689056 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.689110 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.689154 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:53:39 crc kubenswrapper[4967]: E1011 03:53:39.690181 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:40.19016953 +0000 UTC m=+148.153378463 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.690822 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.694635 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.765008 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.791334 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:39 crc kubenswrapper[4967]: E1011 03:53:39.791512 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:40.291485257 +0000 UTC m=+148.254694200 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.791563 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.791630 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.791771 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:39 crc kubenswrapper[4967]: E1011 03:53:39.792192 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:40.292182475 +0000 UTC m=+148.255391408 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.796586 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.801316 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.892356 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:39 crc kubenswrapper[4967]: E1011 03:53:39.892661 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:40.392631828 +0000 UTC m=+148.355840781 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:39 crc kubenswrapper[4967]: I1011 03:53:39.993882 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:39 crc kubenswrapper[4967]: E1011 03:53:39.994224 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:40.494212291 +0000 UTC m=+148.457421224 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.057157 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.073642 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.094647 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:40 crc kubenswrapper[4967]: E1011 03:53:40.095021 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:40.595006323 +0000 UTC m=+148.558215256 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.196037 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:40 crc kubenswrapper[4967]: E1011 03:53:40.196311 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:40.696299917 +0000 UTC m=+148.659508850 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:40 crc kubenswrapper[4967]: W1011 03:53:40.196583 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5fe485a1_e14f_4c09_b5b9_f252bc42b7e8.slice/crio-43ef8e894c63a2b82e2c8bbe34167ec371e6852f6b97045caf35734188bd74d2 WatchSource:0}: Error finding container 43ef8e894c63a2b82e2c8bbe34167ec371e6852f6b97045caf35734188bd74d2: Status 404 returned error can't find the container with id 43ef8e894c63a2b82e2c8bbe34167ec371e6852f6b97045caf35734188bd74d2 Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.261786 4967 patch_prober.go:28] interesting pod/router-default-5444994796-2fvpq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 11 03:53:40 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Oct 11 03:53:40 crc kubenswrapper[4967]: [+]process-running ok Oct 11 03:53:40 crc kubenswrapper[4967]: healthz check failed Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.262136 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2fvpq" podUID="0761866e-d6ab-40c9-a776-8c7fb2683fad" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.306685 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:40 crc kubenswrapper[4967]: E1011 03:53:40.307122 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:40.807105742 +0000 UTC m=+148.770314665 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.330362 4967 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-wnjpj container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.22:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.330429 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnjpj" podUID="e6b311ab-c933-4950-9104-8071d75d77ac" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.22:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.336330 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-dnnwg" event={"ID":"e1a3f43f-63fe-4816-8414-2d7e4c63be57","Type":"ContainerStarted","Data":"041f0e8712057e4d5a02cb5359565899af46dbfca1cc6eef6fb5380b8787e8d5"} Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.336360 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-dnnwg" event={"ID":"e1a3f43f-63fe-4816-8414-2d7e4c63be57","Type":"ContainerStarted","Data":"318b3247795b5c039e33b1a1a81000622b7985c22766d2a731a8cc5cff6bbd3f"} Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.344896 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"43ef8e894c63a2b82e2c8bbe34167ec371e6852f6b97045caf35734188bd74d2"} Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.370403 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-49rxs" Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.411675 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:40 crc kubenswrapper[4967]: E1011 03:53:40.429747 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:40.92973175 +0000 UTC m=+148.892940683 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.491847 4967 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.510090 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-799qw"] Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.511341 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-799qw" Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.512882 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:40 crc kubenswrapper[4967]: E1011 03:53:40.513011 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:41.012991992 +0000 UTC m=+148.976200925 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.513156 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.513196 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-ws6r9" Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.513217 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:40 crc kubenswrapper[4967]: E1011 03:53:40.513593 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:41.013581898 +0000 UTC m=+148.976790831 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.523819 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-799qw"] Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.614692 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.615236 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fx4sl\" (UniqueName: \"kubernetes.io/projected/d05dc46c-e2ec-4c4b-bfed-dd509f93a5df-kube-api-access-fx4sl\") pod \"community-operators-799qw\" (UID: \"d05dc46c-e2ec-4c4b-bfed-dd509f93a5df\") " pod="openshift-marketplace/community-operators-799qw" Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.615292 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d05dc46c-e2ec-4c4b-bfed-dd509f93a5df-catalog-content\") pod \"community-operators-799qw\" (UID: \"d05dc46c-e2ec-4c4b-bfed-dd509f93a5df\") " pod="openshift-marketplace/community-operators-799qw" Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.615313 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d05dc46c-e2ec-4c4b-bfed-dd509f93a5df-utilities\") pod \"community-operators-799qw\" (UID: \"d05dc46c-e2ec-4c4b-bfed-dd509f93a5df\") " pod="openshift-marketplace/community-operators-799qw" Oct 11 03:53:40 crc kubenswrapper[4967]: E1011 03:53:40.616048 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:41.116032706 +0000 UTC m=+149.079241639 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.692556 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zbm5d"] Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.693485 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zbm5d" Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.695659 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.701541 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zbm5d"] Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.717004 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d05dc46c-e2ec-4c4b-bfed-dd509f93a5df-catalog-content\") pod \"community-operators-799qw\" (UID: \"d05dc46c-e2ec-4c4b-bfed-dd509f93a5df\") " pod="openshift-marketplace/community-operators-799qw" Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.717033 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d05dc46c-e2ec-4c4b-bfed-dd509f93a5df-utilities\") pod \"community-operators-799qw\" (UID: \"d05dc46c-e2ec-4c4b-bfed-dd509f93a5df\") " pod="openshift-marketplace/community-operators-799qw" Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.717119 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.717171 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fx4sl\" (UniqueName: \"kubernetes.io/projected/d05dc46c-e2ec-4c4b-bfed-dd509f93a5df-kube-api-access-fx4sl\") pod \"community-operators-799qw\" (UID: \"d05dc46c-e2ec-4c4b-bfed-dd509f93a5df\") " pod="openshift-marketplace/community-operators-799qw" Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.717885 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d05dc46c-e2ec-4c4b-bfed-dd509f93a5df-catalog-content\") pod \"community-operators-799qw\" (UID: \"d05dc46c-e2ec-4c4b-bfed-dd509f93a5df\") " pod="openshift-marketplace/community-operators-799qw" Oct 11 03:53:40 crc kubenswrapper[4967]: E1011 03:53:40.717957 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:41.217941287 +0000 UTC m=+149.181150210 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.717973 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d05dc46c-e2ec-4c4b-bfed-dd509f93a5df-utilities\") pod \"community-operators-799qw\" (UID: \"d05dc46c-e2ec-4c4b-bfed-dd509f93a5df\") " pod="openshift-marketplace/community-operators-799qw" Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.750233 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fx4sl\" (UniqueName: \"kubernetes.io/projected/d05dc46c-e2ec-4c4b-bfed-dd509f93a5df-kube-api-access-fx4sl\") pod \"community-operators-799qw\" (UID: \"d05dc46c-e2ec-4c4b-bfed-dd509f93a5df\") " pod="openshift-marketplace/community-operators-799qw" Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.817697 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.817905 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/130a69d7-5216-4017-9146-1940a1d58210-catalog-content\") pod \"certified-operators-zbm5d\" (UID: \"130a69d7-5216-4017-9146-1940a1d58210\") " pod="openshift-marketplace/certified-operators-zbm5d" Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.817932 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsqz9\" (UniqueName: \"kubernetes.io/projected/130a69d7-5216-4017-9146-1940a1d58210-kube-api-access-vsqz9\") pod \"certified-operators-zbm5d\" (UID: \"130a69d7-5216-4017-9146-1940a1d58210\") " pod="openshift-marketplace/certified-operators-zbm5d" Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.817972 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/130a69d7-5216-4017-9146-1940a1d58210-utilities\") pod \"certified-operators-zbm5d\" (UID: \"130a69d7-5216-4017-9146-1940a1d58210\") " pod="openshift-marketplace/certified-operators-zbm5d" Oct 11 03:53:40 crc kubenswrapper[4967]: E1011 03:53:40.818115 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2025-10-11 03:53:41.318098461 +0000 UTC m=+149.281307394 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.879328 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-799qw" Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.889811 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-9zqq7"] Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.891037 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9zqq7" Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.899264 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9zqq7"] Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.919629 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/130a69d7-5216-4017-9146-1940a1d58210-utilities\") pod \"certified-operators-zbm5d\" (UID: \"130a69d7-5216-4017-9146-1940a1d58210\") " pod="openshift-marketplace/certified-operators-zbm5d" Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.919714 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.919781 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/130a69d7-5216-4017-9146-1940a1d58210-catalog-content\") pod \"certified-operators-zbm5d\" (UID: \"130a69d7-5216-4017-9146-1940a1d58210\") " pod="openshift-marketplace/certified-operators-zbm5d" Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.919808 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsqz9\" (UniqueName: \"kubernetes.io/projected/130a69d7-5216-4017-9146-1940a1d58210-kube-api-access-vsqz9\") pod \"certified-operators-zbm5d\" (UID: \"130a69d7-5216-4017-9146-1940a1d58210\") " pod="openshift-marketplace/certified-operators-zbm5d" Oct 11 03:53:40 crc kubenswrapper[4967]: E1011 03:53:40.920175 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2025-10-11 03:53:41.420159597 +0000 UTC m=+149.383368530 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-hk2cx" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.920380 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/130a69d7-5216-4017-9146-1940a1d58210-utilities\") pod \"certified-operators-zbm5d\" (UID: \"130a69d7-5216-4017-9146-1940a1d58210\") " pod="openshift-marketplace/certified-operators-zbm5d" Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.920400 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/130a69d7-5216-4017-9146-1940a1d58210-catalog-content\") pod \"certified-operators-zbm5d\" (UID: \"130a69d7-5216-4017-9146-1940a1d58210\") " pod="openshift-marketplace/certified-operators-zbm5d" Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.943885 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsqz9\" (UniqueName: \"kubernetes.io/projected/130a69d7-5216-4017-9146-1940a1d58210-kube-api-access-vsqz9\") pod \"certified-operators-zbm5d\" (UID: \"130a69d7-5216-4017-9146-1940a1d58210\") " pod="openshift-marketplace/certified-operators-zbm5d" Oct 11 03:53:40 crc kubenswrapper[4967]: I1011 03:53:40.998623 4967 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2025-10-11T03:53:40.491868416Z","Handler":null,"Name":""} Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.001144 4967 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.001178 4967 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.004365 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zbm5d" Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.021408 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.021686 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9698\" (UniqueName: \"kubernetes.io/projected/e0f97fb5-caf9-44e8-b8f0-14e290c823e5-kube-api-access-h9698\") pod \"community-operators-9zqq7\" (UID: \"e0f97fb5-caf9-44e8-b8f0-14e290c823e5\") " pod="openshift-marketplace/community-operators-9zqq7" Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.021805 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0f97fb5-caf9-44e8-b8f0-14e290c823e5-catalog-content\") pod \"community-operators-9zqq7\" (UID: \"e0f97fb5-caf9-44e8-b8f0-14e290c823e5\") " pod="openshift-marketplace/community-operators-9zqq7" Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.021863 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0f97fb5-caf9-44e8-b8f0-14e290c823e5-utilities\") pod \"community-operators-9zqq7\" (UID: \"e0f97fb5-caf9-44e8-b8f0-14e290c823e5\") " pod="openshift-marketplace/community-operators-9zqq7" Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.026905 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.093734 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-7mg7t"] Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.094646 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7mg7t" Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.105270 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7mg7t"] Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.122819 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0f97fb5-caf9-44e8-b8f0-14e290c823e5-catalog-content\") pod \"community-operators-9zqq7\" (UID: \"e0f97fb5-caf9-44e8-b8f0-14e290c823e5\") " pod="openshift-marketplace/community-operators-9zqq7" Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.122868 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0f97fb5-caf9-44e8-b8f0-14e290c823e5-utilities\") pod \"community-operators-9zqq7\" (UID: \"e0f97fb5-caf9-44e8-b8f0-14e290c823e5\") " pod="openshift-marketplace/community-operators-9zqq7" Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.122910 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9698\" (UniqueName: \"kubernetes.io/projected/e0f97fb5-caf9-44e8-b8f0-14e290c823e5-kube-api-access-h9698\") pod \"community-operators-9zqq7\" (UID: \"e0f97fb5-caf9-44e8-b8f0-14e290c823e5\") " pod="openshift-marketplace/community-operators-9zqq7" Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.122995 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.123773 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0f97fb5-caf9-44e8-b8f0-14e290c823e5-catalog-content\") pod \"community-operators-9zqq7\" (UID: \"e0f97fb5-caf9-44e8-b8f0-14e290c823e5\") " pod="openshift-marketplace/community-operators-9zqq7" Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.124052 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0f97fb5-caf9-44e8-b8f0-14e290c823e5-utilities\") pod \"community-operators-9zqq7\" (UID: \"e0f97fb5-caf9-44e8-b8f0-14e290c823e5\") " pod="openshift-marketplace/community-operators-9zqq7" Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.128889 4967 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.128947 4967 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.147507 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9698\" (UniqueName: \"kubernetes.io/projected/e0f97fb5-caf9-44e8-b8f0-14e290c823e5-kube-api-access-h9698\") pod \"community-operators-9zqq7\" (UID: \"e0f97fb5-caf9-44e8-b8f0-14e290c823e5\") " pod="openshift-marketplace/community-operators-9zqq7" Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.158180 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-799qw"] Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.204842 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-hk2cx\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.216688 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9zqq7" Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.224037 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w98lc\" (UniqueName: \"kubernetes.io/projected/572eea97-2918-420e-821b-b23b212191b3-kube-api-access-w98lc\") pod \"certified-operators-7mg7t\" (UID: \"572eea97-2918-420e-821b-b23b212191b3\") " pod="openshift-marketplace/certified-operators-7mg7t" Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.224124 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/572eea97-2918-420e-821b-b23b212191b3-utilities\") pod \"certified-operators-7mg7t\" (UID: \"572eea97-2918-420e-821b-b23b212191b3\") " pod="openshift-marketplace/certified-operators-7mg7t" Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.224205 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/572eea97-2918-420e-821b-b23b212191b3-catalog-content\") pod \"certified-operators-7mg7t\" (UID: \"572eea97-2918-420e-821b-b23b212191b3\") " pod="openshift-marketplace/certified-operators-7mg7t" Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.253455 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zbm5d"] Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.261280 4967 patch_prober.go:28] interesting pod/router-default-5444994796-2fvpq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 11 03:53:41 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Oct 11 03:53:41 crc kubenswrapper[4967]: [+]process-running ok Oct 11 03:53:41 crc kubenswrapper[4967]: healthz check failed Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.261377 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2fvpq" podUID="0761866e-d6ab-40c9-a776-8c7fb2683fad" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.325624 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w98lc\" (UniqueName: \"kubernetes.io/projected/572eea97-2918-420e-821b-b23b212191b3-kube-api-access-w98lc\") pod \"certified-operators-7mg7t\" (UID: \"572eea97-2918-420e-821b-b23b212191b3\") " pod="openshift-marketplace/certified-operators-7mg7t" Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.325670 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/572eea97-2918-420e-821b-b23b212191b3-utilities\") pod \"certified-operators-7mg7t\" (UID: \"572eea97-2918-420e-821b-b23b212191b3\") " pod="openshift-marketplace/certified-operators-7mg7t" Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.325736 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/572eea97-2918-420e-821b-b23b212191b3-catalog-content\") pod \"certified-operators-7mg7t\" (UID: \"572eea97-2918-420e-821b-b23b212191b3\") " pod="openshift-marketplace/certified-operators-7mg7t" Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.326274 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/572eea97-2918-420e-821b-b23b212191b3-catalog-content\") pod \"certified-operators-7mg7t\" (UID: \"572eea97-2918-420e-821b-b23b212191b3\") " pod="openshift-marketplace/certified-operators-7mg7t" Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.326638 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/572eea97-2918-420e-821b-b23b212191b3-utilities\") pod \"certified-operators-7mg7t\" (UID: \"572eea97-2918-420e-821b-b23b212191b3\") " pod="openshift-marketplace/certified-operators-7mg7t" Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.345791 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w98lc\" (UniqueName: \"kubernetes.io/projected/572eea97-2918-420e-821b-b23b212191b3-kube-api-access-w98lc\") pod \"certified-operators-7mg7t\" (UID: \"572eea97-2918-420e-821b-b23b212191b3\") " pod="openshift-marketplace/certified-operators-7mg7t" Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.360552 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zbm5d" event={"ID":"130a69d7-5216-4017-9146-1940a1d58210","Type":"ContainerStarted","Data":"bad7f074e36a83b3f576de8abd25928f38f125d938fcc2613a2fcb75936b8110"} Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.361897 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"a663151fbf523f598397b6921902d5ad3357f1985fe2afb3a36cdc9d73a4a685"} Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.361929 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"9333ae2d261206ff850feae298a204702acd0d8afa4481163ffb130bb48b265f"} Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.373496 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"51484e7807272685b8cc3344d15f0ec31f5445900371d8cee01b79823529de28"} Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.373564 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"21499b1657d5f2d01db7215045e9c2f6f02a2611cbbeec615c39954ee915f5e0"} Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.376565 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-dnnwg" event={"ID":"e1a3f43f-63fe-4816-8414-2d7e4c63be57","Type":"ContainerStarted","Data":"d0f13411b9949727bdc57e2a77d25476ae3c52399078e1b66a743aa871b634ba"} Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.379747 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"b19c814c9dcced67f4cd3febf9627bbd0a5ff440ef0b490c9716c043e8a9d881"} Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.384284 4967 generic.go:334] "Generic (PLEG): container finished" podID="d05dc46c-e2ec-4c4b-bfed-dd509f93a5df" containerID="00c6be927f7455bc22fdd520a0c4f146bbe7bb1cc95debb3637831fdb5ea1f13" exitCode=0 Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.384487 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-799qw" event={"ID":"d05dc46c-e2ec-4c4b-bfed-dd509f93a5df","Type":"ContainerDied","Data":"00c6be927f7455bc22fdd520a0c4f146bbe7bb1cc95debb3637831fdb5ea1f13"} Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.384550 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-799qw" event={"ID":"d05dc46c-e2ec-4c4b-bfed-dd509f93a5df","Type":"ContainerStarted","Data":"5c224038aef95047b2222fca8bd4225e68df6b467265ec804bb5c595b0947c38"} Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.394337 4967 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.409881 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7mg7t" Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.414238 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-dnnwg" podStartSLOduration=10.414213114 podStartE2EDuration="10.414213114s" podCreationTimestamp="2025-10-11 03:53:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:41.411121879 +0000 UTC m=+149.374330812" watchObservedRunningTime="2025-10-11 03:53:41.414213114 +0000 UTC m=+149.377422047" Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.461774 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.624184 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-9zqq7"] Oct 11 03:53:41 crc kubenswrapper[4967]: W1011 03:53:41.649830 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode0f97fb5_caf9_44e8_b8f0_14e290c823e5.slice/crio-d797f1cb54aa8b409649030ba6a343e0aa4a55f066577741fbe6ab6670253626 WatchSource:0}: Error finding container d797f1cb54aa8b409649030ba6a343e0aa4a55f066577741fbe6ab6670253626: Status 404 returned error can't find the container with id d797f1cb54aa8b409649030ba6a343e0aa4a55f066577741fbe6ab6670253626 Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.696011 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-7mg7t"] Oct 11 03:53:41 crc kubenswrapper[4967]: W1011 03:53:41.722843 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod572eea97_2918_420e_821b_b23b212191b3.slice/crio-27391a31217d116cad7699689f7c00db4c64d53b5cae2704aa55b8453f26fc12 WatchSource:0}: Error finding container 27391a31217d116cad7699689f7c00db4c64d53b5cae2704aa55b8453f26fc12: Status 404 returned error can't find the container with id 27391a31217d116cad7699689f7c00db4c64d53b5cae2704aa55b8453f26fc12 Oct 11 03:53:41 crc kubenswrapper[4967]: I1011 03:53:41.783983 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hk2cx"] Oct 11 03:53:42 crc kubenswrapper[4967]: I1011 03:53:42.260380 4967 patch_prober.go:28] interesting pod/router-default-5444994796-2fvpq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 11 03:53:42 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Oct 11 03:53:42 crc kubenswrapper[4967]: [+]process-running ok Oct 11 03:53:42 crc kubenswrapper[4967]: healthz check failed Oct 11 03:53:42 crc kubenswrapper[4967]: I1011 03:53:42.260444 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2fvpq" podUID="0761866e-d6ab-40c9-a776-8c7fb2683fad" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 11 03:53:42 crc kubenswrapper[4967]: I1011 03:53:42.410236 4967 generic.go:334] "Generic (PLEG): container finished" podID="e0f97fb5-caf9-44e8-b8f0-14e290c823e5" containerID="e46797a464f8e7c83e21369642d58ba2d23b135eecfbd710402a9b0d0f894d5e" exitCode=0 Oct 11 03:53:42 crc kubenswrapper[4967]: I1011 03:53:42.410342 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zqq7" event={"ID":"e0f97fb5-caf9-44e8-b8f0-14e290c823e5","Type":"ContainerDied","Data":"e46797a464f8e7c83e21369642d58ba2d23b135eecfbd710402a9b0d0f894d5e"} Oct 11 03:53:42 crc kubenswrapper[4967]: I1011 03:53:42.410410 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zqq7" event={"ID":"e0f97fb5-caf9-44e8-b8f0-14e290c823e5","Type":"ContainerStarted","Data":"d797f1cb54aa8b409649030ba6a343e0aa4a55f066577741fbe6ab6670253626"} Oct 11 03:53:42 crc kubenswrapper[4967]: I1011 03:53:42.411939 4967 generic.go:334] "Generic (PLEG): container finished" podID="130a69d7-5216-4017-9146-1940a1d58210" containerID="6bb2124bf15e4b330f16e40c088c5045b674c6c278cb382497c812d7d5281f1d" exitCode=0 Oct 11 03:53:42 crc kubenswrapper[4967]: I1011 03:53:42.412585 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zbm5d" event={"ID":"130a69d7-5216-4017-9146-1940a1d58210","Type":"ContainerDied","Data":"6bb2124bf15e4b330f16e40c088c5045b674c6c278cb382497c812d7d5281f1d"} Oct 11 03:53:42 crc kubenswrapper[4967]: I1011 03:53:42.418787 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" event={"ID":"6b6ccb8c-d307-430b-9232-70497214c0bf","Type":"ContainerStarted","Data":"dfdb4b8a671e8fd016f27b7f407b2bbae9a0e9ba57d72096c8669b6263dffa3a"} Oct 11 03:53:42 crc kubenswrapper[4967]: I1011 03:53:42.418826 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" event={"ID":"6b6ccb8c-d307-430b-9232-70497214c0bf","Type":"ContainerStarted","Data":"162a69a1228fdca679e50de3ce6c6a391f937a475b638b6c90bc6d2d053c5a20"} Oct 11 03:53:42 crc kubenswrapper[4967]: I1011 03:53:42.418933 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:53:42 crc kubenswrapper[4967]: I1011 03:53:42.423294 4967 generic.go:334] "Generic (PLEG): container finished" podID="572eea97-2918-420e-821b-b23b212191b3" containerID="2e7282c9d89eeca9f55c23abe1df8837544427b20d113343591f4ec93626e386" exitCode=0 Oct 11 03:53:42 crc kubenswrapper[4967]: I1011 03:53:42.423502 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mg7t" event={"ID":"572eea97-2918-420e-821b-b23b212191b3","Type":"ContainerDied","Data":"2e7282c9d89eeca9f55c23abe1df8837544427b20d113343591f4ec93626e386"} Oct 11 03:53:42 crc kubenswrapper[4967]: I1011 03:53:42.423551 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mg7t" event={"ID":"572eea97-2918-420e-821b-b23b212191b3","Type":"ContainerStarted","Data":"27391a31217d116cad7699689f7c00db4c64d53b5cae2704aa55b8453f26fc12"} Oct 11 03:53:42 crc kubenswrapper[4967]: I1011 03:53:42.449420 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" podStartSLOduration=129.449402233 podStartE2EDuration="2m9.449402233s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:53:42.446186765 +0000 UTC m=+150.409395698" watchObservedRunningTime="2025-10-11 03:53:42.449402233 +0000 UTC m=+150.412611166" Oct 11 03:53:42 crc kubenswrapper[4967]: I1011 03:53:42.691205 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-xjxhr"] Oct 11 03:53:42 crc kubenswrapper[4967]: I1011 03:53:42.692367 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xjxhr" Oct 11 03:53:42 crc kubenswrapper[4967]: I1011 03:53:42.696468 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 11 03:53:42 crc kubenswrapper[4967]: I1011 03:53:42.700735 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xjxhr"] Oct 11 03:53:42 crc kubenswrapper[4967]: I1011 03:53:42.765042 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac-catalog-content\") pod \"redhat-marketplace-xjxhr\" (UID: \"f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac\") " pod="openshift-marketplace/redhat-marketplace-xjxhr" Oct 11 03:53:42 crc kubenswrapper[4967]: I1011 03:53:42.765126 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lws4d\" (UniqueName: \"kubernetes.io/projected/f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac-kube-api-access-lws4d\") pod \"redhat-marketplace-xjxhr\" (UID: \"f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac\") " pod="openshift-marketplace/redhat-marketplace-xjxhr" Oct 11 03:53:42 crc kubenswrapper[4967]: I1011 03:53:42.765183 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac-utilities\") pod \"redhat-marketplace-xjxhr\" (UID: \"f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac\") " pod="openshift-marketplace/redhat-marketplace-xjxhr" Oct 11 03:53:42 crc kubenswrapper[4967]: I1011 03:53:42.822356 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Oct 11 03:53:42 crc kubenswrapper[4967]: I1011 03:53:42.866756 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lws4d\" (UniqueName: \"kubernetes.io/projected/f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac-kube-api-access-lws4d\") pod \"redhat-marketplace-xjxhr\" (UID: \"f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac\") " pod="openshift-marketplace/redhat-marketplace-xjxhr" Oct 11 03:53:42 crc kubenswrapper[4967]: I1011 03:53:42.873372 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac-utilities\") pod \"redhat-marketplace-xjxhr\" (UID: \"f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac\") " pod="openshift-marketplace/redhat-marketplace-xjxhr" Oct 11 03:53:42 crc kubenswrapper[4967]: I1011 03:53:42.873433 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac-catalog-content\") pod \"redhat-marketplace-xjxhr\" (UID: \"f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac\") " pod="openshift-marketplace/redhat-marketplace-xjxhr" Oct 11 03:53:42 crc kubenswrapper[4967]: I1011 03:53:42.874038 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac-catalog-content\") pod \"redhat-marketplace-xjxhr\" (UID: \"f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac\") " pod="openshift-marketplace/redhat-marketplace-xjxhr" Oct 11 03:53:42 crc kubenswrapper[4967]: I1011 03:53:42.874342 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac-utilities\") pod \"redhat-marketplace-xjxhr\" (UID: \"f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac\") " pod="openshift-marketplace/redhat-marketplace-xjxhr" Oct 11 03:53:42 crc kubenswrapper[4967]: I1011 03:53:42.888055 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lws4d\" (UniqueName: \"kubernetes.io/projected/f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac-kube-api-access-lws4d\") pod \"redhat-marketplace-xjxhr\" (UID: \"f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac\") " pod="openshift-marketplace/redhat-marketplace-xjxhr" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.053046 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xjxhr" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.094548 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-kjgbj"] Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.095782 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kjgbj" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.101922 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kjgbj"] Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.142421 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.143060 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.146717 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.146934 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.152989 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.176646 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5af71eaf-029a-4a8a-a563-e17c6eaeb6ce-utilities\") pod \"redhat-marketplace-kjgbj\" (UID: \"5af71eaf-029a-4a8a-a563-e17c6eaeb6ce\") " pod="openshift-marketplace/redhat-marketplace-kjgbj" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.176702 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-czvv7\" (UniqueName: \"kubernetes.io/projected/5af71eaf-029a-4a8a-a563-e17c6eaeb6ce-kube-api-access-czvv7\") pod \"redhat-marketplace-kjgbj\" (UID: \"5af71eaf-029a-4a8a-a563-e17c6eaeb6ce\") " pod="openshift-marketplace/redhat-marketplace-kjgbj" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.176781 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5af71eaf-029a-4a8a-a563-e17c6eaeb6ce-catalog-content\") pod \"redhat-marketplace-kjgbj\" (UID: \"5af71eaf-029a-4a8a-a563-e17c6eaeb6ce\") " pod="openshift-marketplace/redhat-marketplace-kjgbj" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.259853 4967 patch_prober.go:28] interesting pod/router-default-5444994796-2fvpq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 11 03:53:43 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Oct 11 03:53:43 crc kubenswrapper[4967]: [+]process-running ok Oct 11 03:53:43 crc kubenswrapper[4967]: healthz check failed Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.260114 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2fvpq" podUID="0761866e-d6ab-40c9-a776-8c7fb2683fad" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.279522 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/57630488-3eda-4f83-93c6-7c9c5c07a109-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"57630488-3eda-4f83-93c6-7c9c5c07a109\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.279611 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5af71eaf-029a-4a8a-a563-e17c6eaeb6ce-catalog-content\") pod \"redhat-marketplace-kjgbj\" (UID: \"5af71eaf-029a-4a8a-a563-e17c6eaeb6ce\") " pod="openshift-marketplace/redhat-marketplace-kjgbj" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.279686 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5af71eaf-029a-4a8a-a563-e17c6eaeb6ce-utilities\") pod \"redhat-marketplace-kjgbj\" (UID: \"5af71eaf-029a-4a8a-a563-e17c6eaeb6ce\") " pod="openshift-marketplace/redhat-marketplace-kjgbj" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.279728 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-czvv7\" (UniqueName: \"kubernetes.io/projected/5af71eaf-029a-4a8a-a563-e17c6eaeb6ce-kube-api-access-czvv7\") pod \"redhat-marketplace-kjgbj\" (UID: \"5af71eaf-029a-4a8a-a563-e17c6eaeb6ce\") " pod="openshift-marketplace/redhat-marketplace-kjgbj" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.280568 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5af71eaf-029a-4a8a-a563-e17c6eaeb6ce-utilities\") pod \"redhat-marketplace-kjgbj\" (UID: \"5af71eaf-029a-4a8a-a563-e17c6eaeb6ce\") " pod="openshift-marketplace/redhat-marketplace-kjgbj" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.280613 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/57630488-3eda-4f83-93c6-7c9c5c07a109-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"57630488-3eda-4f83-93c6-7c9c5c07a109\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.280676 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5af71eaf-029a-4a8a-a563-e17c6eaeb6ce-catalog-content\") pod \"redhat-marketplace-kjgbj\" (UID: \"5af71eaf-029a-4a8a-a563-e17c6eaeb6ce\") " pod="openshift-marketplace/redhat-marketplace-kjgbj" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.283615 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-xjxhr"] Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.296033 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-czvv7\" (UniqueName: \"kubernetes.io/projected/5af71eaf-029a-4a8a-a563-e17c6eaeb6ce-kube-api-access-czvv7\") pod \"redhat-marketplace-kjgbj\" (UID: \"5af71eaf-029a-4a8a-a563-e17c6eaeb6ce\") " pod="openshift-marketplace/redhat-marketplace-kjgbj" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.382030 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/57630488-3eda-4f83-93c6-7c9c5c07a109-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"57630488-3eda-4f83-93c6-7c9c5c07a109\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.382092 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/57630488-3eda-4f83-93c6-7c9c5c07a109-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"57630488-3eda-4f83-93c6-7c9c5c07a109\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.382370 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/57630488-3eda-4f83-93c6-7c9c5c07a109-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"57630488-3eda-4f83-93c6-7c9c5c07a109\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.399486 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/57630488-3eda-4f83-93c6-7c9c5c07a109-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"57630488-3eda-4f83-93c6-7c9c5c07a109\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.431301 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kjgbj" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.431778 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xjxhr" event={"ID":"f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac","Type":"ContainerStarted","Data":"855379c72f6d636d494e86e891bca308bddf0eb428a87fd239f2ec8b8afa528f"} Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.434417 4967 generic.go:334] "Generic (PLEG): container finished" podID="c506a84b-4ea6-477e-88de-fceec3a95f31" containerID="cc575196dfaaf2394c082a5d499f0ffaae1a90e9adfdc0626c04af2ad9f0fcba" exitCode=0 Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.434506 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-kdpld" event={"ID":"c506a84b-4ea6-477e-88de-fceec3a95f31","Type":"ContainerDied","Data":"cc575196dfaaf2394c082a5d499f0ffaae1a90e9adfdc0626c04af2ad9f0fcba"} Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.459659 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.577180 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.577644 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.586521 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.693863 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vr9z8"] Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.694999 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vr9z8" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.696542 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-kjgbj"] Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.697292 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.710347 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vr9z8"] Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.790856 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tj978\" (UniqueName: \"kubernetes.io/projected/f70cf6aa-6c00-4507-98c3-4f385cc57522-kube-api-access-tj978\") pod \"redhat-operators-vr9z8\" (UID: \"f70cf6aa-6c00-4507-98c3-4f385cc57522\") " pod="openshift-marketplace/redhat-operators-vr9z8" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.790918 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f70cf6aa-6c00-4507-98c3-4f385cc57522-catalog-content\") pod \"redhat-operators-vr9z8\" (UID: \"f70cf6aa-6c00-4507-98c3-4f385cc57522\") " pod="openshift-marketplace/redhat-operators-vr9z8" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.790991 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f70cf6aa-6c00-4507-98c3-4f385cc57522-utilities\") pod \"redhat-operators-vr9z8\" (UID: \"f70cf6aa-6c00-4507-98c3-4f385cc57522\") " pod="openshift-marketplace/redhat-operators-vr9z8" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.810281 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-csl6r" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.810337 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-csl6r" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.822635 4967 patch_prober.go:28] interesting pod/console-f9d7485db-csl6r container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" start-of-body= Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.822686 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-csl6r" podUID="5e20a512-8f34-4d21-be45-e749f7a6cc87" containerName="console" probeResult="failure" output="Get \"https://10.217.0.16:8443/health\": dial tcp 10.217.0.16:8443: connect: connection refused" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.891689 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f70cf6aa-6c00-4507-98c3-4f385cc57522-catalog-content\") pod \"redhat-operators-vr9z8\" (UID: \"f70cf6aa-6c00-4507-98c3-4f385cc57522\") " pod="openshift-marketplace/redhat-operators-vr9z8" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.891850 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f70cf6aa-6c00-4507-98c3-4f385cc57522-utilities\") pod \"redhat-operators-vr9z8\" (UID: \"f70cf6aa-6c00-4507-98c3-4f385cc57522\") " pod="openshift-marketplace/redhat-operators-vr9z8" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.892050 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tj978\" (UniqueName: \"kubernetes.io/projected/f70cf6aa-6c00-4507-98c3-4f385cc57522-kube-api-access-tj978\") pod \"redhat-operators-vr9z8\" (UID: \"f70cf6aa-6c00-4507-98c3-4f385cc57522\") " pod="openshift-marketplace/redhat-operators-vr9z8" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.892644 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f70cf6aa-6c00-4507-98c3-4f385cc57522-catalog-content\") pod \"redhat-operators-vr9z8\" (UID: \"f70cf6aa-6c00-4507-98c3-4f385cc57522\") " pod="openshift-marketplace/redhat-operators-vr9z8" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.892798 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f70cf6aa-6c00-4507-98c3-4f385cc57522-utilities\") pod \"redhat-operators-vr9z8\" (UID: \"f70cf6aa-6c00-4507-98c3-4f385cc57522\") " pod="openshift-marketplace/redhat-operators-vr9z8" Oct 11 03:53:43 crc kubenswrapper[4967]: E1011 03:53:43.906032 4967 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5af71eaf_029a_4a8a_a563_e17c6eaeb6ce.slice/crio-conmon-20e97cc389813cc7ef1b654e4e60744f0e3feca44823afa1f128ad06861cbf70.scope\": RecentStats: unable to find data in memory cache]" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.916608 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tj978\" (UniqueName: \"kubernetes.io/projected/f70cf6aa-6c00-4507-98c3-4f385cc57522-kube-api-access-tj978\") pod \"redhat-operators-vr9z8\" (UID: \"f70cf6aa-6c00-4507-98c3-4f385cc57522\") " pod="openshift-marketplace/redhat-operators-vr9z8" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.984225 4967 patch_prober.go:28] interesting pod/downloads-7954f5f757-dtj88 container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.984266 4967 patch_prober.go:28] interesting pod/downloads-7954f5f757-dtj88 container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" start-of-body= Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.984278 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-dtj88" podUID="d6b9a450-fd86-47a3-9c0e-f61aa7a24377" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Oct 11 03:53:43 crc kubenswrapper[4967]: I1011 03:53:43.984318 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-dtj88" podUID="d6b9a450-fd86-47a3-9c0e-f61aa7a24377" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.10:8080/\": dial tcp 10.217.0.10:8080: connect: connection refused" Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.022252 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vr9z8" Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.051531 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.088985 4967 patch_prober.go:28] interesting pod/machine-config-daemon-sgrv2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.089049 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.094886 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-qjt9g"] Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.096011 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qjt9g" Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.104311 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qjt9g"] Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.164811 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.165458 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.168545 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.168750 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.171787 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.199722 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68f026ee-3e09-46ae-b4af-a92ec0980e21-catalog-content\") pod \"redhat-operators-qjt9g\" (UID: \"68f026ee-3e09-46ae-b4af-a92ec0980e21\") " pod="openshift-marketplace/redhat-operators-qjt9g" Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.199922 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pnjgb\" (UniqueName: \"kubernetes.io/projected/68f026ee-3e09-46ae-b4af-a92ec0980e21-kube-api-access-pnjgb\") pod \"redhat-operators-qjt9g\" (UID: \"68f026ee-3e09-46ae-b4af-a92ec0980e21\") " pod="openshift-marketplace/redhat-operators-qjt9g" Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.199984 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68f026ee-3e09-46ae-b4af-a92ec0980e21-utilities\") pod \"redhat-operators-qjt9g\" (UID: \"68f026ee-3e09-46ae-b4af-a92ec0980e21\") " pod="openshift-marketplace/redhat-operators-qjt9g" Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.257408 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-2fvpq" Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.260308 4967 patch_prober.go:28] interesting pod/router-default-5444994796-2fvpq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 11 03:53:44 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Oct 11 03:53:44 crc kubenswrapper[4967]: [+]process-running ok Oct 11 03:53:44 crc kubenswrapper[4967]: healthz check failed Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.260538 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2fvpq" podUID="0761866e-d6ab-40c9-a776-8c7fb2683fad" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.283228 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vr9z8"] Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.301784 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pnjgb\" (UniqueName: \"kubernetes.io/projected/68f026ee-3e09-46ae-b4af-a92ec0980e21-kube-api-access-pnjgb\") pod \"redhat-operators-qjt9g\" (UID: \"68f026ee-3e09-46ae-b4af-a92ec0980e21\") " pod="openshift-marketplace/redhat-operators-qjt9g" Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.301835 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68f026ee-3e09-46ae-b4af-a92ec0980e21-utilities\") pod \"redhat-operators-qjt9g\" (UID: \"68f026ee-3e09-46ae-b4af-a92ec0980e21\") " pod="openshift-marketplace/redhat-operators-qjt9g" Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.301892 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/46038ef8-bdeb-4ff7-a5b4-bd09512ee26f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"46038ef8-bdeb-4ff7-a5b4-bd09512ee26f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.301908 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68f026ee-3e09-46ae-b4af-a92ec0980e21-catalog-content\") pod \"redhat-operators-qjt9g\" (UID: \"68f026ee-3e09-46ae-b4af-a92ec0980e21\") " pod="openshift-marketplace/redhat-operators-qjt9g" Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.301975 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/46038ef8-bdeb-4ff7-a5b4-bd09512ee26f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"46038ef8-bdeb-4ff7-a5b4-bd09512ee26f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.303959 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68f026ee-3e09-46ae-b4af-a92ec0980e21-utilities\") pod \"redhat-operators-qjt9g\" (UID: \"68f026ee-3e09-46ae-b4af-a92ec0980e21\") " pod="openshift-marketplace/redhat-operators-qjt9g" Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.304264 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68f026ee-3e09-46ae-b4af-a92ec0980e21-catalog-content\") pod \"redhat-operators-qjt9g\" (UID: \"68f026ee-3e09-46ae-b4af-a92ec0980e21\") " pod="openshift-marketplace/redhat-operators-qjt9g" Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.319693 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pnjgb\" (UniqueName: \"kubernetes.io/projected/68f026ee-3e09-46ae-b4af-a92ec0980e21-kube-api-access-pnjgb\") pod \"redhat-operators-qjt9g\" (UID: \"68f026ee-3e09-46ae-b4af-a92ec0980e21\") " pod="openshift-marketplace/redhat-operators-qjt9g" Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.403478 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/46038ef8-bdeb-4ff7-a5b4-bd09512ee26f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"46038ef8-bdeb-4ff7-a5b4-bd09512ee26f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.403571 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/46038ef8-bdeb-4ff7-a5b4-bd09512ee26f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"46038ef8-bdeb-4ff7-a5b4-bd09512ee26f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.403799 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/46038ef8-bdeb-4ff7-a5b4-bd09512ee26f-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"46038ef8-bdeb-4ff7-a5b4-bd09512ee26f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.413355 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qjt9g" Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.420212 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/46038ef8-bdeb-4ff7-a5b4-bd09512ee26f-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"46038ef8-bdeb-4ff7-a5b4-bd09512ee26f\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.442456 4967 generic.go:334] "Generic (PLEG): container finished" podID="f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac" containerID="10943eaca91a949c0d167198c27ed265fa1318ce4219cab5bd17308a8aa32577" exitCode=0 Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.442526 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xjxhr" event={"ID":"f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac","Type":"ContainerDied","Data":"10943eaca91a949c0d167198c27ed265fa1318ce4219cab5bd17308a8aa32577"} Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.447453 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"57630488-3eda-4f83-93c6-7c9c5c07a109","Type":"ContainerStarted","Data":"52a813459078b31e38be6032d5141076c2d906d7c55ec4b09af5cf23fbcc3be4"} Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.451713 4967 generic.go:334] "Generic (PLEG): container finished" podID="5af71eaf-029a-4a8a-a563-e17c6eaeb6ce" containerID="20e97cc389813cc7ef1b654e4e60744f0e3feca44823afa1f128ad06861cbf70" exitCode=0 Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.451753 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kjgbj" event={"ID":"5af71eaf-029a-4a8a-a563-e17c6eaeb6ce","Type":"ContainerDied","Data":"20e97cc389813cc7ef1b654e4e60744f0e3feca44823afa1f128ad06861cbf70"} Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.451805 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kjgbj" event={"ID":"5af71eaf-029a-4a8a-a563-e17c6eaeb6ce","Type":"ContainerStarted","Data":"e3fd2b144cb8db240e09647f99d9cd2190c589ef969d2b51c5595ae1f491c7cd"} Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.456345 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vr9z8" event={"ID":"f70cf6aa-6c00-4507-98c3-4f385cc57522","Type":"ContainerStarted","Data":"f6d5363b33e25ec5859961d37df45e033ef6e51d22836921080fe09b32dabd13"} Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.456374 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vr9z8" event={"ID":"f70cf6aa-6c00-4507-98c3-4f385cc57522","Type":"ContainerStarted","Data":"bab94f0f0c3b481e24f11c67cc7e2a4a657321e134eec71be13d3d99bd92dc43"} Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.460870 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-nxdtr" Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.487938 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.844630 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-kdpld" Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.856353 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-wnjpj" Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.923014 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fwwhz\" (UniqueName: \"kubernetes.io/projected/c506a84b-4ea6-477e-88de-fceec3a95f31-kube-api-access-fwwhz\") pod \"c506a84b-4ea6-477e-88de-fceec3a95f31\" (UID: \"c506a84b-4ea6-477e-88de-fceec3a95f31\") " Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.923133 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c506a84b-4ea6-477e-88de-fceec3a95f31-secret-volume\") pod \"c506a84b-4ea6-477e-88de-fceec3a95f31\" (UID: \"c506a84b-4ea6-477e-88de-fceec3a95f31\") " Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.923230 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c506a84b-4ea6-477e-88de-fceec3a95f31-config-volume\") pod \"c506a84b-4ea6-477e-88de-fceec3a95f31\" (UID: \"c506a84b-4ea6-477e-88de-fceec3a95f31\") " Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.925140 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c506a84b-4ea6-477e-88de-fceec3a95f31-config-volume" (OuterVolumeSpecName: "config-volume") pod "c506a84b-4ea6-477e-88de-fceec3a95f31" (UID: "c506a84b-4ea6-477e-88de-fceec3a95f31"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.929492 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c506a84b-4ea6-477e-88de-fceec3a95f31-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "c506a84b-4ea6-477e-88de-fceec3a95f31" (UID: "c506a84b-4ea6-477e-88de-fceec3a95f31"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:53:44 crc kubenswrapper[4967]: I1011 03:53:44.929821 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c506a84b-4ea6-477e-88de-fceec3a95f31-kube-api-access-fwwhz" (OuterVolumeSpecName: "kube-api-access-fwwhz") pod "c506a84b-4ea6-477e-88de-fceec3a95f31" (UID: "c506a84b-4ea6-477e-88de-fceec3a95f31"). InnerVolumeSpecName "kube-api-access-fwwhz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:53:45 crc kubenswrapper[4967]: I1011 03:53:45.013990 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-qjt9g"] Oct 11 03:53:45 crc kubenswrapper[4967]: I1011 03:53:45.016585 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Oct 11 03:53:45 crc kubenswrapper[4967]: I1011 03:53:45.024629 4967 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/c506a84b-4ea6-477e-88de-fceec3a95f31-config-volume\") on node \"crc\" DevicePath \"\"" Oct 11 03:53:45 crc kubenswrapper[4967]: I1011 03:53:45.024656 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fwwhz\" (UniqueName: \"kubernetes.io/projected/c506a84b-4ea6-477e-88de-fceec3a95f31-kube-api-access-fwwhz\") on node \"crc\" DevicePath \"\"" Oct 11 03:53:45 crc kubenswrapper[4967]: I1011 03:53:45.024668 4967 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/c506a84b-4ea6-477e-88de-fceec3a95f31-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 11 03:53:45 crc kubenswrapper[4967]: W1011 03:53:45.058326 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod68f026ee_3e09_46ae_b4af_a92ec0980e21.slice/crio-9a111b86a37a5d7aed63f95cd51f62a411a9ba264ce6ad318a7ef04c26deb7fb WatchSource:0}: Error finding container 9a111b86a37a5d7aed63f95cd51f62a411a9ba264ce6ad318a7ef04c26deb7fb: Status 404 returned error can't find the container with id 9a111b86a37a5d7aed63f95cd51f62a411a9ba264ce6ad318a7ef04c26deb7fb Oct 11 03:53:45 crc kubenswrapper[4967]: W1011 03:53:45.062239 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod46038ef8_bdeb_4ff7_a5b4_bd09512ee26f.slice/crio-d6dc1d36b84d1fd1ff5e0dfd9d21ea53bdf09f264a89925136dc25a64c18bc12 WatchSource:0}: Error finding container d6dc1d36b84d1fd1ff5e0dfd9d21ea53bdf09f264a89925136dc25a64c18bc12: Status 404 returned error can't find the container with id d6dc1d36b84d1fd1ff5e0dfd9d21ea53bdf09f264a89925136dc25a64c18bc12 Oct 11 03:53:45 crc kubenswrapper[4967]: I1011 03:53:45.261057 4967 patch_prober.go:28] interesting pod/router-default-5444994796-2fvpq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 11 03:53:45 crc kubenswrapper[4967]: [-]has-synced failed: reason withheld Oct 11 03:53:45 crc kubenswrapper[4967]: [+]process-running ok Oct 11 03:53:45 crc kubenswrapper[4967]: healthz check failed Oct 11 03:53:45 crc kubenswrapper[4967]: I1011 03:53:45.261126 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2fvpq" podUID="0761866e-d6ab-40c9-a776-8c7fb2683fad" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 11 03:53:45 crc kubenswrapper[4967]: I1011 03:53:45.467812 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"46038ef8-bdeb-4ff7-a5b4-bd09512ee26f","Type":"ContainerStarted","Data":"d6dc1d36b84d1fd1ff5e0dfd9d21ea53bdf09f264a89925136dc25a64c18bc12"} Oct 11 03:53:45 crc kubenswrapper[4967]: I1011 03:53:45.479918 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-kdpld" Oct 11 03:53:45 crc kubenswrapper[4967]: I1011 03:53:45.479909 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335905-kdpld" event={"ID":"c506a84b-4ea6-477e-88de-fceec3a95f31","Type":"ContainerDied","Data":"4c084fa9b84e40f8ef355f65935253cecd1d31951d86121a4cfbd6bc4b98453a"} Oct 11 03:53:45 crc kubenswrapper[4967]: I1011 03:53:45.480407 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="4c084fa9b84e40f8ef355f65935253cecd1d31951d86121a4cfbd6bc4b98453a" Oct 11 03:53:45 crc kubenswrapper[4967]: I1011 03:53:45.489301 4967 generic.go:334] "Generic (PLEG): container finished" podID="57630488-3eda-4f83-93c6-7c9c5c07a109" containerID="6d4b518d83fdc07e08677df4380b02bcef2d99972528f72ba652da1ca537ea9c" exitCode=0 Oct 11 03:53:45 crc kubenswrapper[4967]: I1011 03:53:45.489734 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"57630488-3eda-4f83-93c6-7c9c5c07a109","Type":"ContainerDied","Data":"6d4b518d83fdc07e08677df4380b02bcef2d99972528f72ba652da1ca537ea9c"} Oct 11 03:53:45 crc kubenswrapper[4967]: I1011 03:53:45.500349 4967 generic.go:334] "Generic (PLEG): container finished" podID="68f026ee-3e09-46ae-b4af-a92ec0980e21" containerID="cb8806e2076616eb108ed0bd53a787cbd88a3e1126a1e89ec97a25e5290ee7c7" exitCode=0 Oct 11 03:53:45 crc kubenswrapper[4967]: I1011 03:53:45.500659 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qjt9g" event={"ID":"68f026ee-3e09-46ae-b4af-a92ec0980e21","Type":"ContainerDied","Data":"cb8806e2076616eb108ed0bd53a787cbd88a3e1126a1e89ec97a25e5290ee7c7"} Oct 11 03:53:45 crc kubenswrapper[4967]: I1011 03:53:45.500727 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qjt9g" event={"ID":"68f026ee-3e09-46ae-b4af-a92ec0980e21","Type":"ContainerStarted","Data":"9a111b86a37a5d7aed63f95cd51f62a411a9ba264ce6ad318a7ef04c26deb7fb"} Oct 11 03:53:45 crc kubenswrapper[4967]: I1011 03:53:45.504734 4967 generic.go:334] "Generic (PLEG): container finished" podID="f70cf6aa-6c00-4507-98c3-4f385cc57522" containerID="f6d5363b33e25ec5859961d37df45e033ef6e51d22836921080fe09b32dabd13" exitCode=0 Oct 11 03:53:45 crc kubenswrapper[4967]: I1011 03:53:45.505397 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vr9z8" event={"ID":"f70cf6aa-6c00-4507-98c3-4f385cc57522","Type":"ContainerDied","Data":"f6d5363b33e25ec5859961d37df45e033ef6e51d22836921080fe09b32dabd13"} Oct 11 03:53:46 crc kubenswrapper[4967]: I1011 03:53:46.260880 4967 patch_prober.go:28] interesting pod/router-default-5444994796-2fvpq container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Oct 11 03:53:46 crc kubenswrapper[4967]: [+]has-synced ok Oct 11 03:53:46 crc kubenswrapper[4967]: [+]process-running ok Oct 11 03:53:46 crc kubenswrapper[4967]: healthz check failed Oct 11 03:53:46 crc kubenswrapper[4967]: I1011 03:53:46.261396 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-2fvpq" podUID="0761866e-d6ab-40c9-a776-8c7fb2683fad" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Oct 11 03:53:46 crc kubenswrapper[4967]: I1011 03:53:46.514247 4967 generic.go:334] "Generic (PLEG): container finished" podID="46038ef8-bdeb-4ff7-a5b4-bd09512ee26f" containerID="34d1cb1364674c20a19d06928b9dfd96d16191b8fbec1de76d03be1500bca506" exitCode=0 Oct 11 03:53:46 crc kubenswrapper[4967]: I1011 03:53:46.514298 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"46038ef8-bdeb-4ff7-a5b4-bd09512ee26f","Type":"ContainerDied","Data":"34d1cb1364674c20a19d06928b9dfd96d16191b8fbec1de76d03be1500bca506"} Oct 11 03:53:47 crc kubenswrapper[4967]: I1011 03:53:47.262203 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-2fvpq" Oct 11 03:53:47 crc kubenswrapper[4967]: I1011 03:53:47.264229 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-2fvpq" Oct 11 03:53:49 crc kubenswrapper[4967]: I1011 03:53:49.637787 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-d8x8n" Oct 11 03:53:50 crc kubenswrapper[4967]: I1011 03:53:50.058644 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:53:53 crc kubenswrapper[4967]: I1011 03:53:53.846575 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 11 03:53:53 crc kubenswrapper[4967]: I1011 03:53:53.853536 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 11 03:53:53 crc kubenswrapper[4967]: I1011 03:53:53.858938 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-csl6r" Oct 11 03:53:53 crc kubenswrapper[4967]: I1011 03:53:53.865210 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-csl6r" Oct 11 03:53:53 crc kubenswrapper[4967]: I1011 03:53:53.961014 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/46038ef8-bdeb-4ff7-a5b4-bd09512ee26f-kube-api-access\") pod \"46038ef8-bdeb-4ff7-a5b4-bd09512ee26f\" (UID: \"46038ef8-bdeb-4ff7-a5b4-bd09512ee26f\") " Oct 11 03:53:53 crc kubenswrapper[4967]: I1011 03:53:53.961168 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/57630488-3eda-4f83-93c6-7c9c5c07a109-kube-api-access\") pod \"57630488-3eda-4f83-93c6-7c9c5c07a109\" (UID: \"57630488-3eda-4f83-93c6-7c9c5c07a109\") " Oct 11 03:53:53 crc kubenswrapper[4967]: I1011 03:53:53.961215 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/46038ef8-bdeb-4ff7-a5b4-bd09512ee26f-kubelet-dir\") pod \"46038ef8-bdeb-4ff7-a5b4-bd09512ee26f\" (UID: \"46038ef8-bdeb-4ff7-a5b4-bd09512ee26f\") " Oct 11 03:53:53 crc kubenswrapper[4967]: I1011 03:53:53.961246 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/57630488-3eda-4f83-93c6-7c9c5c07a109-kubelet-dir\") pod \"57630488-3eda-4f83-93c6-7c9c5c07a109\" (UID: \"57630488-3eda-4f83-93c6-7c9c5c07a109\") " Oct 11 03:53:53 crc kubenswrapper[4967]: I1011 03:53:53.961308 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/46038ef8-bdeb-4ff7-a5b4-bd09512ee26f-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "46038ef8-bdeb-4ff7-a5b4-bd09512ee26f" (UID: "46038ef8-bdeb-4ff7-a5b4-bd09512ee26f"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:53:53 crc kubenswrapper[4967]: I1011 03:53:53.961452 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/57630488-3eda-4f83-93c6-7c9c5c07a109-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "57630488-3eda-4f83-93c6-7c9c5c07a109" (UID: "57630488-3eda-4f83-93c6-7c9c5c07a109"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:53:53 crc kubenswrapper[4967]: I1011 03:53:53.961695 4967 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/46038ef8-bdeb-4ff7-a5b4-bd09512ee26f-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 11 03:53:53 crc kubenswrapper[4967]: I1011 03:53:53.961722 4967 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/57630488-3eda-4f83-93c6-7c9c5c07a109-kubelet-dir\") on node \"crc\" DevicePath \"\"" Oct 11 03:53:53 crc kubenswrapper[4967]: I1011 03:53:53.966731 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/46038ef8-bdeb-4ff7-a5b4-bd09512ee26f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "46038ef8-bdeb-4ff7-a5b4-bd09512ee26f" (UID: "46038ef8-bdeb-4ff7-a5b4-bd09512ee26f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:53:53 crc kubenswrapper[4967]: I1011 03:53:53.967143 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57630488-3eda-4f83-93c6-7c9c5c07a109-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "57630488-3eda-4f83-93c6-7c9c5c07a109" (UID: "57630488-3eda-4f83-93c6-7c9c5c07a109"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:53:53 crc kubenswrapper[4967]: I1011 03:53:53.991133 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-dtj88" Oct 11 03:53:54 crc kubenswrapper[4967]: I1011 03:53:54.062479 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/57630488-3eda-4f83-93c6-7c9c5c07a109-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 11 03:53:54 crc kubenswrapper[4967]: I1011 03:53:54.062518 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/46038ef8-bdeb-4ff7-a5b4-bd09512ee26f-kube-api-access\") on node \"crc\" DevicePath \"\"" Oct 11 03:53:54 crc kubenswrapper[4967]: I1011 03:53:54.588334 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"46038ef8-bdeb-4ff7-a5b4-bd09512ee26f","Type":"ContainerDied","Data":"d6dc1d36b84d1fd1ff5e0dfd9d21ea53bdf09f264a89925136dc25a64c18bc12"} Oct 11 03:53:54 crc kubenswrapper[4967]: I1011 03:53:54.588705 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d6dc1d36b84d1fd1ff5e0dfd9d21ea53bdf09f264a89925136dc25a64c18bc12" Oct 11 03:53:54 crc kubenswrapper[4967]: I1011 03:53:54.588408 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Oct 11 03:53:54 crc kubenswrapper[4967]: I1011 03:53:54.593497 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Oct 11 03:53:54 crc kubenswrapper[4967]: I1011 03:53:54.593487 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"57630488-3eda-4f83-93c6-7c9c5c07a109","Type":"ContainerDied","Data":"52a813459078b31e38be6032d5141076c2d906d7c55ec4b09af5cf23fbcc3be4"} Oct 11 03:53:54 crc kubenswrapper[4967]: I1011 03:53:54.593557 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52a813459078b31e38be6032d5141076c2d906d7c55ec4b09af5cf23fbcc3be4" Oct 11 03:53:56 crc kubenswrapper[4967]: I1011 03:53:56.086970 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d22c953c-0315-406e-9c5d-f1759837750e-metrics-certs\") pod \"network-metrics-daemon-p24x8\" (UID: \"d22c953c-0315-406e-9c5d-f1759837750e\") " pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:53:56 crc kubenswrapper[4967]: I1011 03:53:56.101856 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/d22c953c-0315-406e-9c5d-f1759837750e-metrics-certs\") pod \"network-metrics-daemon-p24x8\" (UID: \"d22c953c-0315-406e-9c5d-f1759837750e\") " pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:53:56 crc kubenswrapper[4967]: I1011 03:53:56.242919 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-p24x8" Oct 11 03:54:01 crc kubenswrapper[4967]: I1011 03:54:01.468712 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:54:10 crc kubenswrapper[4967]: I1011 03:54:10.066332 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Oct 11 03:54:12 crc kubenswrapper[4967]: E1011 03:54:12.623641 4967 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Oct 11 03:54:12 crc kubenswrapper[4967]: E1011 03:54:12.623934 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-fx4sl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-799qw_openshift-marketplace(d05dc46c-e2ec-4c4b-bfed-dd509f93a5df): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 11 03:54:12 crc kubenswrapper[4967]: E1011 03:54:12.625138 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-799qw" podUID="d05dc46c-e2ec-4c4b-bfed-dd509f93a5df" Oct 11 03:54:12 crc kubenswrapper[4967]: E1011 03:54:12.852549 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-799qw" podUID="d05dc46c-e2ec-4c4b-bfed-dd509f93a5df" Oct 11 03:54:12 crc kubenswrapper[4967]: E1011 03:54:12.983973 4967 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Oct 11 03:54:12 crc kubenswrapper[4967]: E1011 03:54:12.984177 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-w98lc,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-7mg7t_openshift-marketplace(572eea97-2918-420e-821b-b23b212191b3): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 11 03:54:12 crc kubenswrapper[4967]: E1011 03:54:12.985370 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-7mg7t" podUID="572eea97-2918-420e-821b-b23b212191b3" Oct 11 03:54:12 crc kubenswrapper[4967]: E1011 03:54:12.990845 4967 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 11 03:54:12 crc kubenswrapper[4967]: E1011 03:54:12.990948 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lws4d,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-xjxhr_openshift-marketplace(f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 11 03:54:12 crc kubenswrapper[4967]: E1011 03:54:12.992146 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-xjxhr" podUID="f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac" Oct 11 03:54:13 crc kubenswrapper[4967]: E1011 03:54:13.079382 4967 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/redhat-marketplace-index:v4.18" Oct 11 03:54:13 crc kubenswrapper[4967]: E1011 03:54:13.079653 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/redhat-marketplace-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-czvv7,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod redhat-marketplace-kjgbj_openshift-marketplace(5af71eaf-029a-4a8a-a563-e17c6eaeb6ce): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Oct 11 03:54:13 crc kubenswrapper[4967]: E1011 03:54:13.081122 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/redhat-marketplace-kjgbj" podUID="5af71eaf-029a-4a8a-a563-e17c6eaeb6ce" Oct 11 03:54:14 crc kubenswrapper[4967]: I1011 03:54:14.089163 4967 patch_prober.go:28] interesting pod/machine-config-daemon-sgrv2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:54:14 crc kubenswrapper[4967]: I1011 03:54:14.089242 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:54:14 crc kubenswrapper[4967]: I1011 03:54:14.581432 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-mhnj2" Oct 11 03:54:16 crc kubenswrapper[4967]: E1011 03:54:15.999528 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-xjxhr" podUID="f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac" Oct 11 03:54:16 crc kubenswrapper[4967]: E1011 03:54:15.999649 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"\"" pod="openshift-marketplace/redhat-marketplace-kjgbj" podUID="5af71eaf-029a-4a8a-a563-e17c6eaeb6ce" Oct 11 03:54:16 crc kubenswrapper[4967]: E1011 03:54:16.013679 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-7mg7t" podUID="572eea97-2918-420e-821b-b23b212191b3" Oct 11 03:54:16 crc kubenswrapper[4967]: I1011 03:54:16.447912 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-p24x8"] Oct 11 03:54:16 crc kubenswrapper[4967]: W1011 03:54:16.457573 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd22c953c_0315_406e_9c5d_f1759837750e.slice/crio-dc99229cf42ad1101784dad28d9bd21c3146bed79365d75763bc0e43e3406f77 WatchSource:0}: Error finding container dc99229cf42ad1101784dad28d9bd21c3146bed79365d75763bc0e43e3406f77: Status 404 returned error can't find the container with id dc99229cf42ad1101784dad28d9bd21c3146bed79365d75763bc0e43e3406f77 Oct 11 03:54:16 crc kubenswrapper[4967]: I1011 03:54:16.733810 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qjt9g" event={"ID":"68f026ee-3e09-46ae-b4af-a92ec0980e21","Type":"ContainerStarted","Data":"1044dd1091aa4eceb3ceaa220daa7120934782b795f641ab6108a708179cdfd0"} Oct 11 03:54:16 crc kubenswrapper[4967]: I1011 03:54:16.739684 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vr9z8" event={"ID":"f70cf6aa-6c00-4507-98c3-4f385cc57522","Type":"ContainerStarted","Data":"9eaf587a3d43f9a38d7ddb02adadc6b95b93dfd979b538d7908d39d25729c31e"} Oct 11 03:54:16 crc kubenswrapper[4967]: I1011 03:54:16.742018 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-p24x8" event={"ID":"d22c953c-0315-406e-9c5d-f1759837750e","Type":"ContainerStarted","Data":"e10fe8474fad9a69bca0be8344c5e708b00f1e20b919edbc9d9f69d3295f2b4f"} Oct 11 03:54:16 crc kubenswrapper[4967]: I1011 03:54:16.742057 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-p24x8" event={"ID":"d22c953c-0315-406e-9c5d-f1759837750e","Type":"ContainerStarted","Data":"dc99229cf42ad1101784dad28d9bd21c3146bed79365d75763bc0e43e3406f77"} Oct 11 03:54:16 crc kubenswrapper[4967]: I1011 03:54:16.744044 4967 generic.go:334] "Generic (PLEG): container finished" podID="e0f97fb5-caf9-44e8-b8f0-14e290c823e5" containerID="18e508a983765511a422cc7c0b4b396343c9471746a5972d9b94e5777f0a1ee7" exitCode=0 Oct 11 03:54:16 crc kubenswrapper[4967]: I1011 03:54:16.744183 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zqq7" event={"ID":"e0f97fb5-caf9-44e8-b8f0-14e290c823e5","Type":"ContainerDied","Data":"18e508a983765511a422cc7c0b4b396343c9471746a5972d9b94e5777f0a1ee7"} Oct 11 03:54:16 crc kubenswrapper[4967]: I1011 03:54:16.752412 4967 generic.go:334] "Generic (PLEG): container finished" podID="130a69d7-5216-4017-9146-1940a1d58210" containerID="85a60964d4075d899e009d9edf9be19ba25286ca3247114709d8e187293aae6e" exitCode=0 Oct 11 03:54:16 crc kubenswrapper[4967]: I1011 03:54:16.752463 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zbm5d" event={"ID":"130a69d7-5216-4017-9146-1940a1d58210","Type":"ContainerDied","Data":"85a60964d4075d899e009d9edf9be19ba25286ca3247114709d8e187293aae6e"} Oct 11 03:54:17 crc kubenswrapper[4967]: I1011 03:54:17.762108 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zqq7" event={"ID":"e0f97fb5-caf9-44e8-b8f0-14e290c823e5","Type":"ContainerStarted","Data":"5eb7e0ae6964cb12ebd14cb4635d4488ea1a957143a41c1049aeac4e6424bdab"} Oct 11 03:54:17 crc kubenswrapper[4967]: I1011 03:54:17.764584 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zbm5d" event={"ID":"130a69d7-5216-4017-9146-1940a1d58210","Type":"ContainerStarted","Data":"c6da23a3cca166867af38c21c9c0a7ae952b046c5c4c942e8b9ac1952ad01af4"} Oct 11 03:54:17 crc kubenswrapper[4967]: I1011 03:54:17.768605 4967 generic.go:334] "Generic (PLEG): container finished" podID="68f026ee-3e09-46ae-b4af-a92ec0980e21" containerID="1044dd1091aa4eceb3ceaa220daa7120934782b795f641ab6108a708179cdfd0" exitCode=0 Oct 11 03:54:17 crc kubenswrapper[4967]: I1011 03:54:17.768666 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qjt9g" event={"ID":"68f026ee-3e09-46ae-b4af-a92ec0980e21","Type":"ContainerDied","Data":"1044dd1091aa4eceb3ceaa220daa7120934782b795f641ab6108a708179cdfd0"} Oct 11 03:54:17 crc kubenswrapper[4967]: I1011 03:54:17.771351 4967 generic.go:334] "Generic (PLEG): container finished" podID="f70cf6aa-6c00-4507-98c3-4f385cc57522" containerID="9eaf587a3d43f9a38d7ddb02adadc6b95b93dfd979b538d7908d39d25729c31e" exitCode=0 Oct 11 03:54:17 crc kubenswrapper[4967]: I1011 03:54:17.771428 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vr9z8" event={"ID":"f70cf6aa-6c00-4507-98c3-4f385cc57522","Type":"ContainerDied","Data":"9eaf587a3d43f9a38d7ddb02adadc6b95b93dfd979b538d7908d39d25729c31e"} Oct 11 03:54:17 crc kubenswrapper[4967]: I1011 03:54:17.774151 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-p24x8" event={"ID":"d22c953c-0315-406e-9c5d-f1759837750e","Type":"ContainerStarted","Data":"0776aeaa9f9845ef72a2cff8620c68a6742c18b0d144cdf8856a678ee93d46a1"} Oct 11 03:54:17 crc kubenswrapper[4967]: I1011 03:54:17.812198 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-9zqq7" podStartSLOduration=3.052355672 podStartE2EDuration="37.812175482s" podCreationTimestamp="2025-10-11 03:53:40 +0000 UTC" firstStartedPulling="2025-10-11 03:53:42.412602528 +0000 UTC m=+150.375811461" lastFinishedPulling="2025-10-11 03:54:17.172422338 +0000 UTC m=+185.135631271" observedRunningTime="2025-10-11 03:54:17.786036628 +0000 UTC m=+185.749245601" watchObservedRunningTime="2025-10-11 03:54:17.812175482 +0000 UTC m=+185.775384455" Oct 11 03:54:17 crc kubenswrapper[4967]: I1011 03:54:17.854272 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-p24x8" podStartSLOduration=164.85425315 podStartE2EDuration="2m44.85425315s" podCreationTimestamp="2025-10-11 03:51:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:54:17.831661893 +0000 UTC m=+185.794870846" watchObservedRunningTime="2025-10-11 03:54:17.85425315 +0000 UTC m=+185.817462103" Oct 11 03:54:19 crc kubenswrapper[4967]: I1011 03:54:19.788860 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qjt9g" event={"ID":"68f026ee-3e09-46ae-b4af-a92ec0980e21","Type":"ContainerStarted","Data":"c7820f652cc96e5e0d113453b4c91993fdbae8ab899cb8a7f39fb96ea9f9114d"} Oct 11 03:54:19 crc kubenswrapper[4967]: I1011 03:54:19.791104 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vr9z8" event={"ID":"f70cf6aa-6c00-4507-98c3-4f385cc57522","Type":"ContainerStarted","Data":"eeaf8e1eb112fbeb11b2e4302eba226b06d6e6f1980762e3d027caac5a150ce6"} Oct 11 03:54:19 crc kubenswrapper[4967]: I1011 03:54:19.817474 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zbm5d" podStartSLOduration=5.001013758 podStartE2EDuration="39.817457822s" podCreationTimestamp="2025-10-11 03:53:40 +0000 UTC" firstStartedPulling="2025-10-11 03:53:42.417928144 +0000 UTC m=+150.381137077" lastFinishedPulling="2025-10-11 03:54:17.234372208 +0000 UTC m=+185.197581141" observedRunningTime="2025-10-11 03:54:17.88429649 +0000 UTC m=+185.847505443" watchObservedRunningTime="2025-10-11 03:54:19.817457822 +0000 UTC m=+187.780666755" Oct 11 03:54:19 crc kubenswrapper[4967]: I1011 03:54:19.818193 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-qjt9g" podStartSLOduration=2.405463696 podStartE2EDuration="35.818187922s" podCreationTimestamp="2025-10-11 03:53:44 +0000 UTC" firstStartedPulling="2025-10-11 03:53:45.503283658 +0000 UTC m=+153.466492591" lastFinishedPulling="2025-10-11 03:54:18.916007864 +0000 UTC m=+186.879216817" observedRunningTime="2025-10-11 03:54:19.81590138 +0000 UTC m=+187.779110353" watchObservedRunningTime="2025-10-11 03:54:19.818187922 +0000 UTC m=+187.781396855" Oct 11 03:54:19 crc kubenswrapper[4967]: I1011 03:54:19.843787 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vr9z8" podStartSLOduration=3.69716326 podStartE2EDuration="36.84376555s" podCreationTimestamp="2025-10-11 03:53:43 +0000 UTC" firstStartedPulling="2025-10-11 03:53:45.508327496 +0000 UTC m=+153.471536429" lastFinishedPulling="2025-10-11 03:54:18.654929776 +0000 UTC m=+186.618138719" observedRunningTime="2025-10-11 03:54:19.842557557 +0000 UTC m=+187.805766500" watchObservedRunningTime="2025-10-11 03:54:19.84376555 +0000 UTC m=+187.806974493" Oct 11 03:54:21 crc kubenswrapper[4967]: I1011 03:54:21.005400 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zbm5d" Oct 11 03:54:21 crc kubenswrapper[4967]: I1011 03:54:21.005692 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zbm5d" Oct 11 03:54:21 crc kubenswrapper[4967]: I1011 03:54:21.158880 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zbm5d" Oct 11 03:54:21 crc kubenswrapper[4967]: I1011 03:54:21.217921 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-9zqq7" Oct 11 03:54:21 crc kubenswrapper[4967]: I1011 03:54:21.218333 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-9zqq7" Oct 11 03:54:21 crc kubenswrapper[4967]: I1011 03:54:21.269619 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-9zqq7" Oct 11 03:54:22 crc kubenswrapper[4967]: I1011 03:54:22.866424 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-9zqq7" Oct 11 03:54:24 crc kubenswrapper[4967]: I1011 03:54:24.022776 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vr9z8" Oct 11 03:54:24 crc kubenswrapper[4967]: I1011 03:54:24.023198 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vr9z8" Oct 11 03:54:24 crc kubenswrapper[4967]: I1011 03:54:24.414490 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-qjt9g" Oct 11 03:54:24 crc kubenswrapper[4967]: I1011 03:54:24.414671 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-qjt9g" Oct 11 03:54:24 crc kubenswrapper[4967]: I1011 03:54:24.467206 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-qjt9g" Oct 11 03:54:24 crc kubenswrapper[4967]: I1011 03:54:24.872954 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-qjt9g" Oct 11 03:54:25 crc kubenswrapper[4967]: I1011 03:54:25.106296 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vr9z8" podUID="f70cf6aa-6c00-4507-98c3-4f385cc57522" containerName="registry-server" probeResult="failure" output=< Oct 11 03:54:25 crc kubenswrapper[4967]: timeout: failed to connect service ":50051" within 1s Oct 11 03:54:25 crc kubenswrapper[4967]: > Oct 11 03:54:25 crc kubenswrapper[4967]: I1011 03:54:25.309991 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9zqq7"] Oct 11 03:54:25 crc kubenswrapper[4967]: I1011 03:54:25.310487 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-9zqq7" podUID="e0f97fb5-caf9-44e8-b8f0-14e290c823e5" containerName="registry-server" containerID="cri-o://5eb7e0ae6964cb12ebd14cb4635d4488ea1a957143a41c1049aeac4e6424bdab" gracePeriod=2 Oct 11 03:54:25 crc kubenswrapper[4967]: I1011 03:54:25.827309 4967 generic.go:334] "Generic (PLEG): container finished" podID="e0f97fb5-caf9-44e8-b8f0-14e290c823e5" containerID="5eb7e0ae6964cb12ebd14cb4635d4488ea1a957143a41c1049aeac4e6424bdab" exitCode=0 Oct 11 03:54:25 crc kubenswrapper[4967]: I1011 03:54:25.827378 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zqq7" event={"ID":"e0f97fb5-caf9-44e8-b8f0-14e290c823e5","Type":"ContainerDied","Data":"5eb7e0ae6964cb12ebd14cb4635d4488ea1a957143a41c1049aeac4e6424bdab"} Oct 11 03:54:26 crc kubenswrapper[4967]: I1011 03:54:26.278938 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9zqq7" Oct 11 03:54:26 crc kubenswrapper[4967]: I1011 03:54:26.395029 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9698\" (UniqueName: \"kubernetes.io/projected/e0f97fb5-caf9-44e8-b8f0-14e290c823e5-kube-api-access-h9698\") pod \"e0f97fb5-caf9-44e8-b8f0-14e290c823e5\" (UID: \"e0f97fb5-caf9-44e8-b8f0-14e290c823e5\") " Oct 11 03:54:26 crc kubenswrapper[4967]: I1011 03:54:26.395150 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0f97fb5-caf9-44e8-b8f0-14e290c823e5-catalog-content\") pod \"e0f97fb5-caf9-44e8-b8f0-14e290c823e5\" (UID: \"e0f97fb5-caf9-44e8-b8f0-14e290c823e5\") " Oct 11 03:54:26 crc kubenswrapper[4967]: I1011 03:54:26.395177 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0f97fb5-caf9-44e8-b8f0-14e290c823e5-utilities\") pod \"e0f97fb5-caf9-44e8-b8f0-14e290c823e5\" (UID: \"e0f97fb5-caf9-44e8-b8f0-14e290c823e5\") " Oct 11 03:54:26 crc kubenswrapper[4967]: I1011 03:54:26.397106 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0f97fb5-caf9-44e8-b8f0-14e290c823e5-utilities" (OuterVolumeSpecName: "utilities") pod "e0f97fb5-caf9-44e8-b8f0-14e290c823e5" (UID: "e0f97fb5-caf9-44e8-b8f0-14e290c823e5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:54:26 crc kubenswrapper[4967]: I1011 03:54:26.402835 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e0f97fb5-caf9-44e8-b8f0-14e290c823e5-kube-api-access-h9698" (OuterVolumeSpecName: "kube-api-access-h9698") pod "e0f97fb5-caf9-44e8-b8f0-14e290c823e5" (UID: "e0f97fb5-caf9-44e8-b8f0-14e290c823e5"). InnerVolumeSpecName "kube-api-access-h9698". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:54:26 crc kubenswrapper[4967]: I1011 03:54:26.465002 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e0f97fb5-caf9-44e8-b8f0-14e290c823e5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e0f97fb5-caf9-44e8-b8f0-14e290c823e5" (UID: "e0f97fb5-caf9-44e8-b8f0-14e290c823e5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:54:26 crc kubenswrapper[4967]: I1011 03:54:26.496638 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e0f97fb5-caf9-44e8-b8f0-14e290c823e5-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:54:26 crc kubenswrapper[4967]: I1011 03:54:26.497373 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e0f97fb5-caf9-44e8-b8f0-14e290c823e5-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:54:26 crc kubenswrapper[4967]: I1011 03:54:26.497443 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9698\" (UniqueName: \"kubernetes.io/projected/e0f97fb5-caf9-44e8-b8f0-14e290c823e5-kube-api-access-h9698\") on node \"crc\" DevicePath \"\"" Oct 11 03:54:26 crc kubenswrapper[4967]: I1011 03:54:26.835596 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-9zqq7" Oct 11 03:54:26 crc kubenswrapper[4967]: I1011 03:54:26.835582 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-9zqq7" event={"ID":"e0f97fb5-caf9-44e8-b8f0-14e290c823e5","Type":"ContainerDied","Data":"d797f1cb54aa8b409649030ba6a343e0aa4a55f066577741fbe6ab6670253626"} Oct 11 03:54:26 crc kubenswrapper[4967]: I1011 03:54:26.835766 4967 scope.go:117] "RemoveContainer" containerID="5eb7e0ae6964cb12ebd14cb4635d4488ea1a957143a41c1049aeac4e6424bdab" Oct 11 03:54:26 crc kubenswrapper[4967]: I1011 03:54:26.859280 4967 scope.go:117] "RemoveContainer" containerID="18e508a983765511a422cc7c0b4b396343c9471746a5972d9b94e5777f0a1ee7" Oct 11 03:54:26 crc kubenswrapper[4967]: I1011 03:54:26.887046 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-9zqq7"] Oct 11 03:54:26 crc kubenswrapper[4967]: I1011 03:54:26.894379 4967 scope.go:117] "RemoveContainer" containerID="e46797a464f8e7c83e21369642d58ba2d23b135eecfbd710402a9b0d0f894d5e" Oct 11 03:54:26 crc kubenswrapper[4967]: I1011 03:54:26.896831 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-9zqq7"] Oct 11 03:54:27 crc kubenswrapper[4967]: I1011 03:54:27.111709 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qjt9g"] Oct 11 03:54:27 crc kubenswrapper[4967]: I1011 03:54:27.850734 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-qjt9g" podUID="68f026ee-3e09-46ae-b4af-a92ec0980e21" containerName="registry-server" containerID="cri-o://c7820f652cc96e5e0d113453b4c91993fdbae8ab899cb8a7f39fb96ea9f9114d" gracePeriod=2 Oct 11 03:54:28 crc kubenswrapper[4967]: I1011 03:54:28.831688 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e0f97fb5-caf9-44e8-b8f0-14e290c823e5" path="/var/lib/kubelet/pods/e0f97fb5-caf9-44e8-b8f0-14e290c823e5/volumes" Oct 11 03:54:29 crc kubenswrapper[4967]: I1011 03:54:29.867354 4967 generic.go:334] "Generic (PLEG): container finished" podID="68f026ee-3e09-46ae-b4af-a92ec0980e21" containerID="c7820f652cc96e5e0d113453b4c91993fdbae8ab899cb8a7f39fb96ea9f9114d" exitCode=0 Oct 11 03:54:29 crc kubenswrapper[4967]: I1011 03:54:29.867530 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qjt9g" event={"ID":"68f026ee-3e09-46ae-b4af-a92ec0980e21","Type":"ContainerDied","Data":"c7820f652cc96e5e0d113453b4c91993fdbae8ab899cb8a7f39fb96ea9f9114d"} Oct 11 03:54:30 crc kubenswrapper[4967]: I1011 03:54:30.291529 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qjt9g" Oct 11 03:54:30 crc kubenswrapper[4967]: I1011 03:54:30.467447 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68f026ee-3e09-46ae-b4af-a92ec0980e21-catalog-content\") pod \"68f026ee-3e09-46ae-b4af-a92ec0980e21\" (UID: \"68f026ee-3e09-46ae-b4af-a92ec0980e21\") " Oct 11 03:54:30 crc kubenswrapper[4967]: I1011 03:54:30.467556 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pnjgb\" (UniqueName: \"kubernetes.io/projected/68f026ee-3e09-46ae-b4af-a92ec0980e21-kube-api-access-pnjgb\") pod \"68f026ee-3e09-46ae-b4af-a92ec0980e21\" (UID: \"68f026ee-3e09-46ae-b4af-a92ec0980e21\") " Oct 11 03:54:30 crc kubenswrapper[4967]: I1011 03:54:30.467660 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68f026ee-3e09-46ae-b4af-a92ec0980e21-utilities\") pod \"68f026ee-3e09-46ae-b4af-a92ec0980e21\" (UID: \"68f026ee-3e09-46ae-b4af-a92ec0980e21\") " Oct 11 03:54:30 crc kubenswrapper[4967]: I1011 03:54:30.468444 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68f026ee-3e09-46ae-b4af-a92ec0980e21-utilities" (OuterVolumeSpecName: "utilities") pod "68f026ee-3e09-46ae-b4af-a92ec0980e21" (UID: "68f026ee-3e09-46ae-b4af-a92ec0980e21"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:54:30 crc kubenswrapper[4967]: I1011 03:54:30.476592 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/68f026ee-3e09-46ae-b4af-a92ec0980e21-kube-api-access-pnjgb" (OuterVolumeSpecName: "kube-api-access-pnjgb") pod "68f026ee-3e09-46ae-b4af-a92ec0980e21" (UID: "68f026ee-3e09-46ae-b4af-a92ec0980e21"). InnerVolumeSpecName "kube-api-access-pnjgb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:54:30 crc kubenswrapper[4967]: I1011 03:54:30.569880 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/68f026ee-3e09-46ae-b4af-a92ec0980e21-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:54:30 crc kubenswrapper[4967]: I1011 03:54:30.569934 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pnjgb\" (UniqueName: \"kubernetes.io/projected/68f026ee-3e09-46ae-b4af-a92ec0980e21-kube-api-access-pnjgb\") on node \"crc\" DevicePath \"\"" Oct 11 03:54:30 crc kubenswrapper[4967]: I1011 03:54:30.877271 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-qjt9g" event={"ID":"68f026ee-3e09-46ae-b4af-a92ec0980e21","Type":"ContainerDied","Data":"9a111b86a37a5d7aed63f95cd51f62a411a9ba264ce6ad318a7ef04c26deb7fb"} Oct 11 03:54:30 crc kubenswrapper[4967]: I1011 03:54:30.877324 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-qjt9g" Oct 11 03:54:30 crc kubenswrapper[4967]: I1011 03:54:30.877666 4967 scope.go:117] "RemoveContainer" containerID="c7820f652cc96e5e0d113453b4c91993fdbae8ab899cb8a7f39fb96ea9f9114d" Oct 11 03:54:30 crc kubenswrapper[4967]: I1011 03:54:30.903721 4967 scope.go:117] "RemoveContainer" containerID="1044dd1091aa4eceb3ceaa220daa7120934782b795f641ab6108a708179cdfd0" Oct 11 03:54:30 crc kubenswrapper[4967]: I1011 03:54:30.925792 4967 scope.go:117] "RemoveContainer" containerID="cb8806e2076616eb108ed0bd53a787cbd88a3e1126a1e89ec97a25e5290ee7c7" Oct 11 03:54:31 crc kubenswrapper[4967]: I1011 03:54:31.084652 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zbm5d" Oct 11 03:54:31 crc kubenswrapper[4967]: I1011 03:54:31.087687 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/68f026ee-3e09-46ae-b4af-a92ec0980e21-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "68f026ee-3e09-46ae-b4af-a92ec0980e21" (UID: "68f026ee-3e09-46ae-b4af-a92ec0980e21"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:54:31 crc kubenswrapper[4967]: I1011 03:54:31.183911 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/68f026ee-3e09-46ae-b4af-a92ec0980e21-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:54:31 crc kubenswrapper[4967]: I1011 03:54:31.215225 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-qjt9g"] Oct 11 03:54:31 crc kubenswrapper[4967]: I1011 03:54:31.219767 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-qjt9g"] Oct 11 03:54:32 crc kubenswrapper[4967]: I1011 03:54:32.820031 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="68f026ee-3e09-46ae-b4af-a92ec0980e21" path="/var/lib/kubelet/pods/68f026ee-3e09-46ae-b4af-a92ec0980e21/volumes" Oct 11 03:54:33 crc kubenswrapper[4967]: I1011 03:54:33.900865 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mg7t" event={"ID":"572eea97-2918-420e-821b-b23b212191b3","Type":"ContainerStarted","Data":"a55e60927310d635053c998aa59b6a43be95df3e4f218b645b5a93da6d619887"} Oct 11 03:54:33 crc kubenswrapper[4967]: I1011 03:54:33.906559 4967 generic.go:334] "Generic (PLEG): container finished" podID="5af71eaf-029a-4a8a-a563-e17c6eaeb6ce" containerID="2ba1d6077693b47e4752e058fbef2d44c21858d62809d849593eb1be265da9f5" exitCode=0 Oct 11 03:54:33 crc kubenswrapper[4967]: I1011 03:54:33.906642 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kjgbj" event={"ID":"5af71eaf-029a-4a8a-a563-e17c6eaeb6ce","Type":"ContainerDied","Data":"2ba1d6077693b47e4752e058fbef2d44c21858d62809d849593eb1be265da9f5"} Oct 11 03:54:33 crc kubenswrapper[4967]: I1011 03:54:33.909768 4967 generic.go:334] "Generic (PLEG): container finished" podID="f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac" containerID="4b7afad0f405055b4293e3c4950d487dfe0f0ba03ed20ebf9f7e4fc8041552de" exitCode=0 Oct 11 03:54:33 crc kubenswrapper[4967]: I1011 03:54:33.909810 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xjxhr" event={"ID":"f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac","Type":"ContainerDied","Data":"4b7afad0f405055b4293e3c4950d487dfe0f0ba03ed20ebf9f7e4fc8041552de"} Oct 11 03:54:33 crc kubenswrapper[4967]: I1011 03:54:33.912949 4967 generic.go:334] "Generic (PLEG): container finished" podID="d05dc46c-e2ec-4c4b-bfed-dd509f93a5df" containerID="7961c1c0282827e842a607993d2202e88cc54fbad70c26ebe17d5e0485ec6625" exitCode=0 Oct 11 03:54:33 crc kubenswrapper[4967]: I1011 03:54:33.912973 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-799qw" event={"ID":"d05dc46c-e2ec-4c4b-bfed-dd509f93a5df","Type":"ContainerDied","Data":"7961c1c0282827e842a607993d2202e88cc54fbad70c26ebe17d5e0485ec6625"} Oct 11 03:54:34 crc kubenswrapper[4967]: I1011 03:54:34.072657 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vr9z8" Oct 11 03:54:34 crc kubenswrapper[4967]: I1011 03:54:34.129162 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vr9z8" Oct 11 03:54:34 crc kubenswrapper[4967]: I1011 03:54:34.919214 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xjxhr" event={"ID":"f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac","Type":"ContainerStarted","Data":"0f6ada5b1586532169620a1612c62c2adbf3a4d2e52d6c2f7a6abfc351b4d535"} Oct 11 03:54:34 crc kubenswrapper[4967]: I1011 03:54:34.921679 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-799qw" event={"ID":"d05dc46c-e2ec-4c4b-bfed-dd509f93a5df","Type":"ContainerStarted","Data":"a44900551f8820045090c5a4323ab5056e638679919c4e2e2907861091d4e005"} Oct 11 03:54:34 crc kubenswrapper[4967]: I1011 03:54:34.923637 4967 generic.go:334] "Generic (PLEG): container finished" podID="572eea97-2918-420e-821b-b23b212191b3" containerID="a55e60927310d635053c998aa59b6a43be95df3e4f218b645b5a93da6d619887" exitCode=0 Oct 11 03:54:34 crc kubenswrapper[4967]: I1011 03:54:34.923726 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mg7t" event={"ID":"572eea97-2918-420e-821b-b23b212191b3","Type":"ContainerDied","Data":"a55e60927310d635053c998aa59b6a43be95df3e4f218b645b5a93da6d619887"} Oct 11 03:54:34 crc kubenswrapper[4967]: I1011 03:54:34.929603 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kjgbj" event={"ID":"5af71eaf-029a-4a8a-a563-e17c6eaeb6ce","Type":"ContainerStarted","Data":"4c93f5249cf91038ca483587ed4eae3933a722ccd5b43a32fd3c1a700183b42d"} Oct 11 03:54:34 crc kubenswrapper[4967]: I1011 03:54:34.944560 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-xjxhr" podStartSLOduration=2.76797449 podStartE2EDuration="52.94453917s" podCreationTimestamp="2025-10-11 03:53:42 +0000 UTC" firstStartedPulling="2025-10-11 03:53:44.444324511 +0000 UTC m=+152.407533444" lastFinishedPulling="2025-10-11 03:54:34.620889191 +0000 UTC m=+202.584098124" observedRunningTime="2025-10-11 03:54:34.939895316 +0000 UTC m=+202.903104269" watchObservedRunningTime="2025-10-11 03:54:34.94453917 +0000 UTC m=+202.907748113" Oct 11 03:54:34 crc kubenswrapper[4967]: I1011 03:54:34.991277 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-kjgbj" podStartSLOduration=1.870355511 podStartE2EDuration="51.991253032s" podCreationTimestamp="2025-10-11 03:53:43 +0000 UTC" firstStartedPulling="2025-10-11 03:53:44.45895294 +0000 UTC m=+152.422161873" lastFinishedPulling="2025-10-11 03:54:34.579850461 +0000 UTC m=+202.543059394" observedRunningTime="2025-10-11 03:54:34.986690561 +0000 UTC m=+202.949899494" watchObservedRunningTime="2025-10-11 03:54:34.991253032 +0000 UTC m=+202.954461965" Oct 11 03:54:35 crc kubenswrapper[4967]: I1011 03:54:35.938269 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mg7t" event={"ID":"572eea97-2918-420e-821b-b23b212191b3","Type":"ContainerStarted","Data":"e0a35eddcd970de4002990e9a40230c288dfc572a3f1cbee7fddcb051c359d15"} Oct 11 03:54:35 crc kubenswrapper[4967]: I1011 03:54:35.957815 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-7mg7t" podStartSLOduration=1.974546785 podStartE2EDuration="54.957800882s" podCreationTimestamp="2025-10-11 03:53:41 +0000 UTC" firstStartedPulling="2025-10-11 03:53:42.425659695 +0000 UTC m=+150.388868628" lastFinishedPulling="2025-10-11 03:54:35.408913802 +0000 UTC m=+203.372122725" observedRunningTime="2025-10-11 03:54:35.956883547 +0000 UTC m=+203.920092480" watchObservedRunningTime="2025-10-11 03:54:35.957800882 +0000 UTC m=+203.921009815" Oct 11 03:54:35 crc kubenswrapper[4967]: I1011 03:54:35.959352 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-799qw" podStartSLOduration=3.014294237 podStartE2EDuration="55.959346687s" podCreationTimestamp="2025-10-11 03:53:40 +0000 UTC" firstStartedPulling="2025-10-11 03:53:41.393738525 +0000 UTC m=+149.356947458" lastFinishedPulling="2025-10-11 03:54:34.338790975 +0000 UTC m=+202.301999908" observedRunningTime="2025-10-11 03:54:35.02877819 +0000 UTC m=+202.991987123" watchObservedRunningTime="2025-10-11 03:54:35.959346687 +0000 UTC m=+203.922555620" Oct 11 03:54:40 crc kubenswrapper[4967]: I1011 03:54:40.879732 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-799qw" Oct 11 03:54:40 crc kubenswrapper[4967]: I1011 03:54:40.880710 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-799qw" Oct 11 03:54:40 crc kubenswrapper[4967]: I1011 03:54:40.931037 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-799qw" Oct 11 03:54:40 crc kubenswrapper[4967]: I1011 03:54:40.998544 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-799qw" Oct 11 03:54:41 crc kubenswrapper[4967]: I1011 03:54:41.411013 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-7mg7t" Oct 11 03:54:41 crc kubenswrapper[4967]: I1011 03:54:41.412480 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-7mg7t" Oct 11 03:54:41 crc kubenswrapper[4967]: I1011 03:54:41.457771 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-7mg7t" Oct 11 03:54:42 crc kubenswrapper[4967]: I1011 03:54:42.039960 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-7mg7t" Oct 11 03:54:43 crc kubenswrapper[4967]: I1011 03:54:43.053499 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-xjxhr" Oct 11 03:54:43 crc kubenswrapper[4967]: I1011 03:54:43.053863 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-xjxhr" Oct 11 03:54:43 crc kubenswrapper[4967]: I1011 03:54:43.097600 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-xjxhr" Oct 11 03:54:43 crc kubenswrapper[4967]: I1011 03:54:43.432538 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-kjgbj" Oct 11 03:54:43 crc kubenswrapper[4967]: I1011 03:54:43.432741 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-kjgbj" Oct 11 03:54:43 crc kubenswrapper[4967]: I1011 03:54:43.467248 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-kjgbj" Oct 11 03:54:43 crc kubenswrapper[4967]: I1011 03:54:43.503496 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7mg7t"] Oct 11 03:54:44 crc kubenswrapper[4967]: I1011 03:54:44.011928 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-kjgbj" Oct 11 03:54:44 crc kubenswrapper[4967]: I1011 03:54:44.013963 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-xjxhr" Oct 11 03:54:44 crc kubenswrapper[4967]: I1011 03:54:44.088444 4967 patch_prober.go:28] interesting pod/machine-config-daemon-sgrv2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:54:44 crc kubenswrapper[4967]: I1011 03:54:44.088506 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:54:44 crc kubenswrapper[4967]: I1011 03:54:44.088548 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" Oct 11 03:54:44 crc kubenswrapper[4967]: I1011 03:54:44.089200 4967 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb"} pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 11 03:54:44 crc kubenswrapper[4967]: I1011 03:54:44.089303 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" containerID="cri-o://a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb" gracePeriod=600 Oct 11 03:54:44 crc kubenswrapper[4967]: I1011 03:54:44.982119 4967 generic.go:334] "Generic (PLEG): container finished" podID="97c07678-14be-410c-b61f-498cb49bc960" containerID="a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb" exitCode=0 Oct 11 03:54:44 crc kubenswrapper[4967]: I1011 03:54:44.982206 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" event={"ID":"97c07678-14be-410c-b61f-498cb49bc960","Type":"ContainerDied","Data":"a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb"} Oct 11 03:54:44 crc kubenswrapper[4967]: I1011 03:54:44.983027 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" event={"ID":"97c07678-14be-410c-b61f-498cb49bc960","Type":"ContainerStarted","Data":"7cfd4e9315984debe791c2e9396234b602578eb12f2fc553132988f8ec022205"} Oct 11 03:54:44 crc kubenswrapper[4967]: I1011 03:54:44.983087 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-7mg7t" podUID="572eea97-2918-420e-821b-b23b212191b3" containerName="registry-server" containerID="cri-o://e0a35eddcd970de4002990e9a40230c288dfc572a3f1cbee7fddcb051c359d15" gracePeriod=2 Oct 11 03:54:45 crc kubenswrapper[4967]: I1011 03:54:45.343960 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7mg7t" Oct 11 03:54:45 crc kubenswrapper[4967]: I1011 03:54:45.467616 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/572eea97-2918-420e-821b-b23b212191b3-catalog-content\") pod \"572eea97-2918-420e-821b-b23b212191b3\" (UID: \"572eea97-2918-420e-821b-b23b212191b3\") " Oct 11 03:54:45 crc kubenswrapper[4967]: I1011 03:54:45.467749 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/572eea97-2918-420e-821b-b23b212191b3-utilities\") pod \"572eea97-2918-420e-821b-b23b212191b3\" (UID: \"572eea97-2918-420e-821b-b23b212191b3\") " Oct 11 03:54:45 crc kubenswrapper[4967]: I1011 03:54:45.467772 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w98lc\" (UniqueName: \"kubernetes.io/projected/572eea97-2918-420e-821b-b23b212191b3-kube-api-access-w98lc\") pod \"572eea97-2918-420e-821b-b23b212191b3\" (UID: \"572eea97-2918-420e-821b-b23b212191b3\") " Oct 11 03:54:45 crc kubenswrapper[4967]: I1011 03:54:45.468559 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/572eea97-2918-420e-821b-b23b212191b3-utilities" (OuterVolumeSpecName: "utilities") pod "572eea97-2918-420e-821b-b23b212191b3" (UID: "572eea97-2918-420e-821b-b23b212191b3"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:54:45 crc kubenswrapper[4967]: I1011 03:54:45.474196 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/572eea97-2918-420e-821b-b23b212191b3-kube-api-access-w98lc" (OuterVolumeSpecName: "kube-api-access-w98lc") pod "572eea97-2918-420e-821b-b23b212191b3" (UID: "572eea97-2918-420e-821b-b23b212191b3"). InnerVolumeSpecName "kube-api-access-w98lc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:54:45 crc kubenswrapper[4967]: I1011 03:54:45.569311 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/572eea97-2918-420e-821b-b23b212191b3-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:54:45 crc kubenswrapper[4967]: I1011 03:54:45.569348 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w98lc\" (UniqueName: \"kubernetes.io/projected/572eea97-2918-420e-821b-b23b212191b3-kube-api-access-w98lc\") on node \"crc\" DevicePath \"\"" Oct 11 03:54:45 crc kubenswrapper[4967]: I1011 03:54:45.864676 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/572eea97-2918-420e-821b-b23b212191b3-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "572eea97-2918-420e-821b-b23b212191b3" (UID: "572eea97-2918-420e-821b-b23b212191b3"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:54:45 crc kubenswrapper[4967]: I1011 03:54:45.871716 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/572eea97-2918-420e-821b-b23b212191b3-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:54:45 crc kubenswrapper[4967]: I1011 03:54:45.907792 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kjgbj"] Oct 11 03:54:46 crc kubenswrapper[4967]: I1011 03:54:46.001285 4967 generic.go:334] "Generic (PLEG): container finished" podID="572eea97-2918-420e-821b-b23b212191b3" containerID="e0a35eddcd970de4002990e9a40230c288dfc572a3f1cbee7fddcb051c359d15" exitCode=0 Oct 11 03:54:46 crc kubenswrapper[4967]: I1011 03:54:46.001899 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-7mg7t" Oct 11 03:54:46 crc kubenswrapper[4967]: I1011 03:54:46.002447 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mg7t" event={"ID":"572eea97-2918-420e-821b-b23b212191b3","Type":"ContainerDied","Data":"e0a35eddcd970de4002990e9a40230c288dfc572a3f1cbee7fddcb051c359d15"} Oct 11 03:54:46 crc kubenswrapper[4967]: I1011 03:54:46.002506 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-7mg7t" event={"ID":"572eea97-2918-420e-821b-b23b212191b3","Type":"ContainerDied","Data":"27391a31217d116cad7699689f7c00db4c64d53b5cae2704aa55b8453f26fc12"} Oct 11 03:54:46 crc kubenswrapper[4967]: I1011 03:54:46.002529 4967 scope.go:117] "RemoveContainer" containerID="e0a35eddcd970de4002990e9a40230c288dfc572a3f1cbee7fddcb051c359d15" Oct 11 03:54:46 crc kubenswrapper[4967]: I1011 03:54:46.029267 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-7mg7t"] Oct 11 03:54:46 crc kubenswrapper[4967]: I1011 03:54:46.031534 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-7mg7t"] Oct 11 03:54:46 crc kubenswrapper[4967]: I1011 03:54:46.035886 4967 scope.go:117] "RemoveContainer" containerID="a55e60927310d635053c998aa59b6a43be95df3e4f218b645b5a93da6d619887" Oct 11 03:54:46 crc kubenswrapper[4967]: I1011 03:54:46.054178 4967 scope.go:117] "RemoveContainer" containerID="2e7282c9d89eeca9f55c23abe1df8837544427b20d113343591f4ec93626e386" Oct 11 03:54:46 crc kubenswrapper[4967]: I1011 03:54:46.066464 4967 scope.go:117] "RemoveContainer" containerID="e0a35eddcd970de4002990e9a40230c288dfc572a3f1cbee7fddcb051c359d15" Oct 11 03:54:46 crc kubenswrapper[4967]: E1011 03:54:46.066843 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e0a35eddcd970de4002990e9a40230c288dfc572a3f1cbee7fddcb051c359d15\": container with ID starting with e0a35eddcd970de4002990e9a40230c288dfc572a3f1cbee7fddcb051c359d15 not found: ID does not exist" containerID="e0a35eddcd970de4002990e9a40230c288dfc572a3f1cbee7fddcb051c359d15" Oct 11 03:54:46 crc kubenswrapper[4967]: I1011 03:54:46.066881 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e0a35eddcd970de4002990e9a40230c288dfc572a3f1cbee7fddcb051c359d15"} err="failed to get container status \"e0a35eddcd970de4002990e9a40230c288dfc572a3f1cbee7fddcb051c359d15\": rpc error: code = NotFound desc = could not find container \"e0a35eddcd970de4002990e9a40230c288dfc572a3f1cbee7fddcb051c359d15\": container with ID starting with e0a35eddcd970de4002990e9a40230c288dfc572a3f1cbee7fddcb051c359d15 not found: ID does not exist" Oct 11 03:54:46 crc kubenswrapper[4967]: I1011 03:54:46.066905 4967 scope.go:117] "RemoveContainer" containerID="a55e60927310d635053c998aa59b6a43be95df3e4f218b645b5a93da6d619887" Oct 11 03:54:46 crc kubenswrapper[4967]: E1011 03:54:46.067319 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a55e60927310d635053c998aa59b6a43be95df3e4f218b645b5a93da6d619887\": container with ID starting with a55e60927310d635053c998aa59b6a43be95df3e4f218b645b5a93da6d619887 not found: ID does not exist" containerID="a55e60927310d635053c998aa59b6a43be95df3e4f218b645b5a93da6d619887" Oct 11 03:54:46 crc kubenswrapper[4967]: I1011 03:54:46.067362 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a55e60927310d635053c998aa59b6a43be95df3e4f218b645b5a93da6d619887"} err="failed to get container status \"a55e60927310d635053c998aa59b6a43be95df3e4f218b645b5a93da6d619887\": rpc error: code = NotFound desc = could not find container \"a55e60927310d635053c998aa59b6a43be95df3e4f218b645b5a93da6d619887\": container with ID starting with a55e60927310d635053c998aa59b6a43be95df3e4f218b645b5a93da6d619887 not found: ID does not exist" Oct 11 03:54:46 crc kubenswrapper[4967]: I1011 03:54:46.067391 4967 scope.go:117] "RemoveContainer" containerID="2e7282c9d89eeca9f55c23abe1df8837544427b20d113343591f4ec93626e386" Oct 11 03:54:46 crc kubenswrapper[4967]: E1011 03:54:46.067796 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e7282c9d89eeca9f55c23abe1df8837544427b20d113343591f4ec93626e386\": container with ID starting with 2e7282c9d89eeca9f55c23abe1df8837544427b20d113343591f4ec93626e386 not found: ID does not exist" containerID="2e7282c9d89eeca9f55c23abe1df8837544427b20d113343591f4ec93626e386" Oct 11 03:54:46 crc kubenswrapper[4967]: I1011 03:54:46.067836 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e7282c9d89eeca9f55c23abe1df8837544427b20d113343591f4ec93626e386"} err="failed to get container status \"2e7282c9d89eeca9f55c23abe1df8837544427b20d113343591f4ec93626e386\": rpc error: code = NotFound desc = could not find container \"2e7282c9d89eeca9f55c23abe1df8837544427b20d113343591f4ec93626e386\": container with ID starting with 2e7282c9d89eeca9f55c23abe1df8837544427b20d113343591f4ec93626e386 not found: ID does not exist" Oct 11 03:54:46 crc kubenswrapper[4967]: I1011 03:54:46.821639 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="572eea97-2918-420e-821b-b23b212191b3" path="/var/lib/kubelet/pods/572eea97-2918-420e-821b-b23b212191b3/volumes" Oct 11 03:54:47 crc kubenswrapper[4967]: I1011 03:54:47.007371 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-kjgbj" podUID="5af71eaf-029a-4a8a-a563-e17c6eaeb6ce" containerName="registry-server" containerID="cri-o://4c93f5249cf91038ca483587ed4eae3933a722ccd5b43a32fd3c1a700183b42d" gracePeriod=2 Oct 11 03:54:47 crc kubenswrapper[4967]: I1011 03:54:47.334210 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kjgbj" Oct 11 03:54:47 crc kubenswrapper[4967]: I1011 03:54:47.491609 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5af71eaf-029a-4a8a-a563-e17c6eaeb6ce-catalog-content\") pod \"5af71eaf-029a-4a8a-a563-e17c6eaeb6ce\" (UID: \"5af71eaf-029a-4a8a-a563-e17c6eaeb6ce\") " Oct 11 03:54:47 crc kubenswrapper[4967]: I1011 03:54:47.491695 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-czvv7\" (UniqueName: \"kubernetes.io/projected/5af71eaf-029a-4a8a-a563-e17c6eaeb6ce-kube-api-access-czvv7\") pod \"5af71eaf-029a-4a8a-a563-e17c6eaeb6ce\" (UID: \"5af71eaf-029a-4a8a-a563-e17c6eaeb6ce\") " Oct 11 03:54:47 crc kubenswrapper[4967]: I1011 03:54:47.491781 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5af71eaf-029a-4a8a-a563-e17c6eaeb6ce-utilities\") pod \"5af71eaf-029a-4a8a-a563-e17c6eaeb6ce\" (UID: \"5af71eaf-029a-4a8a-a563-e17c6eaeb6ce\") " Oct 11 03:54:47 crc kubenswrapper[4967]: I1011 03:54:47.492622 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5af71eaf-029a-4a8a-a563-e17c6eaeb6ce-utilities" (OuterVolumeSpecName: "utilities") pod "5af71eaf-029a-4a8a-a563-e17c6eaeb6ce" (UID: "5af71eaf-029a-4a8a-a563-e17c6eaeb6ce"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:54:47 crc kubenswrapper[4967]: I1011 03:54:47.499934 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5af71eaf-029a-4a8a-a563-e17c6eaeb6ce-kube-api-access-czvv7" (OuterVolumeSpecName: "kube-api-access-czvv7") pod "5af71eaf-029a-4a8a-a563-e17c6eaeb6ce" (UID: "5af71eaf-029a-4a8a-a563-e17c6eaeb6ce"). InnerVolumeSpecName "kube-api-access-czvv7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:54:47 crc kubenswrapper[4967]: I1011 03:54:47.504887 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5af71eaf-029a-4a8a-a563-e17c6eaeb6ce-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5af71eaf-029a-4a8a-a563-e17c6eaeb6ce" (UID: "5af71eaf-029a-4a8a-a563-e17c6eaeb6ce"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:54:47 crc kubenswrapper[4967]: I1011 03:54:47.593200 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5af71eaf-029a-4a8a-a563-e17c6eaeb6ce-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:54:47 crc kubenswrapper[4967]: I1011 03:54:47.593229 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-czvv7\" (UniqueName: \"kubernetes.io/projected/5af71eaf-029a-4a8a-a563-e17c6eaeb6ce-kube-api-access-czvv7\") on node \"crc\" DevicePath \"\"" Oct 11 03:54:47 crc kubenswrapper[4967]: I1011 03:54:47.593243 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5af71eaf-029a-4a8a-a563-e17c6eaeb6ce-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:54:48 crc kubenswrapper[4967]: I1011 03:54:48.013446 4967 generic.go:334] "Generic (PLEG): container finished" podID="5af71eaf-029a-4a8a-a563-e17c6eaeb6ce" containerID="4c93f5249cf91038ca483587ed4eae3933a722ccd5b43a32fd3c1a700183b42d" exitCode=0 Oct 11 03:54:48 crc kubenswrapper[4967]: I1011 03:54:48.013487 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kjgbj" event={"ID":"5af71eaf-029a-4a8a-a563-e17c6eaeb6ce","Type":"ContainerDied","Data":"4c93f5249cf91038ca483587ed4eae3933a722ccd5b43a32fd3c1a700183b42d"} Oct 11 03:54:48 crc kubenswrapper[4967]: I1011 03:54:48.013515 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-kjgbj" event={"ID":"5af71eaf-029a-4a8a-a563-e17c6eaeb6ce","Type":"ContainerDied","Data":"e3fd2b144cb8db240e09647f99d9cd2190c589ef969d2b51c5595ae1f491c7cd"} Oct 11 03:54:48 crc kubenswrapper[4967]: I1011 03:54:48.013537 4967 scope.go:117] "RemoveContainer" containerID="4c93f5249cf91038ca483587ed4eae3933a722ccd5b43a32fd3c1a700183b42d" Oct 11 03:54:48 crc kubenswrapper[4967]: I1011 03:54:48.013539 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-kjgbj" Oct 11 03:54:48 crc kubenswrapper[4967]: I1011 03:54:48.033132 4967 scope.go:117] "RemoveContainer" containerID="2ba1d6077693b47e4752e058fbef2d44c21858d62809d849593eb1be265da9f5" Oct 11 03:54:48 crc kubenswrapper[4967]: I1011 03:54:48.042815 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-kjgbj"] Oct 11 03:54:48 crc kubenswrapper[4967]: I1011 03:54:48.045859 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-kjgbj"] Oct 11 03:54:48 crc kubenswrapper[4967]: I1011 03:54:48.068190 4967 scope.go:117] "RemoveContainer" containerID="20e97cc389813cc7ef1b654e4e60744f0e3feca44823afa1f128ad06861cbf70" Oct 11 03:54:48 crc kubenswrapper[4967]: I1011 03:54:48.079920 4967 scope.go:117] "RemoveContainer" containerID="4c93f5249cf91038ca483587ed4eae3933a722ccd5b43a32fd3c1a700183b42d" Oct 11 03:54:48 crc kubenswrapper[4967]: E1011 03:54:48.080437 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4c93f5249cf91038ca483587ed4eae3933a722ccd5b43a32fd3c1a700183b42d\": container with ID starting with 4c93f5249cf91038ca483587ed4eae3933a722ccd5b43a32fd3c1a700183b42d not found: ID does not exist" containerID="4c93f5249cf91038ca483587ed4eae3933a722ccd5b43a32fd3c1a700183b42d" Oct 11 03:54:48 crc kubenswrapper[4967]: I1011 03:54:48.080478 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4c93f5249cf91038ca483587ed4eae3933a722ccd5b43a32fd3c1a700183b42d"} err="failed to get container status \"4c93f5249cf91038ca483587ed4eae3933a722ccd5b43a32fd3c1a700183b42d\": rpc error: code = NotFound desc = could not find container \"4c93f5249cf91038ca483587ed4eae3933a722ccd5b43a32fd3c1a700183b42d\": container with ID starting with 4c93f5249cf91038ca483587ed4eae3933a722ccd5b43a32fd3c1a700183b42d not found: ID does not exist" Oct 11 03:54:48 crc kubenswrapper[4967]: I1011 03:54:48.080511 4967 scope.go:117] "RemoveContainer" containerID="2ba1d6077693b47e4752e058fbef2d44c21858d62809d849593eb1be265da9f5" Oct 11 03:54:48 crc kubenswrapper[4967]: E1011 03:54:48.080791 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2ba1d6077693b47e4752e058fbef2d44c21858d62809d849593eb1be265da9f5\": container with ID starting with 2ba1d6077693b47e4752e058fbef2d44c21858d62809d849593eb1be265da9f5 not found: ID does not exist" containerID="2ba1d6077693b47e4752e058fbef2d44c21858d62809d849593eb1be265da9f5" Oct 11 03:54:48 crc kubenswrapper[4967]: I1011 03:54:48.080826 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2ba1d6077693b47e4752e058fbef2d44c21858d62809d849593eb1be265da9f5"} err="failed to get container status \"2ba1d6077693b47e4752e058fbef2d44c21858d62809d849593eb1be265da9f5\": rpc error: code = NotFound desc = could not find container \"2ba1d6077693b47e4752e058fbef2d44c21858d62809d849593eb1be265da9f5\": container with ID starting with 2ba1d6077693b47e4752e058fbef2d44c21858d62809d849593eb1be265da9f5 not found: ID does not exist" Oct 11 03:54:48 crc kubenswrapper[4967]: I1011 03:54:48.080845 4967 scope.go:117] "RemoveContainer" containerID="20e97cc389813cc7ef1b654e4e60744f0e3feca44823afa1f128ad06861cbf70" Oct 11 03:54:48 crc kubenswrapper[4967]: E1011 03:54:48.081286 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20e97cc389813cc7ef1b654e4e60744f0e3feca44823afa1f128ad06861cbf70\": container with ID starting with 20e97cc389813cc7ef1b654e4e60744f0e3feca44823afa1f128ad06861cbf70 not found: ID does not exist" containerID="20e97cc389813cc7ef1b654e4e60744f0e3feca44823afa1f128ad06861cbf70" Oct 11 03:54:48 crc kubenswrapper[4967]: I1011 03:54:48.081320 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20e97cc389813cc7ef1b654e4e60744f0e3feca44823afa1f128ad06861cbf70"} err="failed to get container status \"20e97cc389813cc7ef1b654e4e60744f0e3feca44823afa1f128ad06861cbf70\": rpc error: code = NotFound desc = could not find container \"20e97cc389813cc7ef1b654e4e60744f0e3feca44823afa1f128ad06861cbf70\": container with ID starting with 20e97cc389813cc7ef1b654e4e60744f0e3feca44823afa1f128ad06861cbf70 not found: ID does not exist" Oct 11 03:54:48 crc kubenswrapper[4967]: I1011 03:54:48.821186 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5af71eaf-029a-4a8a-a563-e17c6eaeb6ce" path="/var/lib/kubelet/pods/5af71eaf-029a-4a8a-a563-e17c6eaeb6ce/volumes" Oct 11 03:54:53 crc kubenswrapper[4967]: I1011 03:54:53.047046 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6sjfd"] Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.084225 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" podUID="0dae89fc-445f-4531-8928-29b334636522" containerName="oauth-openshift" containerID="cri-o://21a94896d91097d19aeb6b46354b56d5640a05428b63933176e0a1178bb5f651" gracePeriod=15 Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.500605 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.560510 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl"] Oct 11 03:55:18 crc kubenswrapper[4967]: E1011 03:55:18.561003 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="57630488-3eda-4f83-93c6-7c9c5c07a109" containerName="pruner" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.561132 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="57630488-3eda-4f83-93c6-7c9c5c07a109" containerName="pruner" Oct 11 03:55:18 crc kubenswrapper[4967]: E1011 03:55:18.561175 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5af71eaf-029a-4a8a-a563-e17c6eaeb6ce" containerName="extract-content" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.561243 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="5af71eaf-029a-4a8a-a563-e17c6eaeb6ce" containerName="extract-content" Oct 11 03:55:18 crc kubenswrapper[4967]: E1011 03:55:18.561264 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0f97fb5-caf9-44e8-b8f0-14e290c823e5" containerName="registry-server" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.561333 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0f97fb5-caf9-44e8-b8f0-14e290c823e5" containerName="registry-server" Oct 11 03:55:18 crc kubenswrapper[4967]: E1011 03:55:18.561356 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5af71eaf-029a-4a8a-a563-e17c6eaeb6ce" containerName="registry-server" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.561374 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="5af71eaf-029a-4a8a-a563-e17c6eaeb6ce" containerName="registry-server" Oct 11 03:55:18 crc kubenswrapper[4967]: E1011 03:55:18.561441 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5af71eaf-029a-4a8a-a563-e17c6eaeb6ce" containerName="extract-utilities" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.561459 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="5af71eaf-029a-4a8a-a563-e17c6eaeb6ce" containerName="extract-utilities" Oct 11 03:55:18 crc kubenswrapper[4967]: E1011 03:55:18.561525 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="572eea97-2918-420e-821b-b23b212191b3" containerName="registry-server" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.561544 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="572eea97-2918-420e-821b-b23b212191b3" containerName="registry-server" Oct 11 03:55:18 crc kubenswrapper[4967]: E1011 03:55:18.561567 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68f026ee-3e09-46ae-b4af-a92ec0980e21" containerName="extract-utilities" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.561584 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="68f026ee-3e09-46ae-b4af-a92ec0980e21" containerName="extract-utilities" Oct 11 03:55:18 crc kubenswrapper[4967]: E1011 03:55:18.561609 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68f026ee-3e09-46ae-b4af-a92ec0980e21" containerName="extract-content" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.561625 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="68f026ee-3e09-46ae-b4af-a92ec0980e21" containerName="extract-content" Oct 11 03:55:18 crc kubenswrapper[4967]: E1011 03:55:18.561647 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="572eea97-2918-420e-821b-b23b212191b3" containerName="extract-utilities" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.561664 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="572eea97-2918-420e-821b-b23b212191b3" containerName="extract-utilities" Oct 11 03:55:18 crc kubenswrapper[4967]: E1011 03:55:18.561690 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c506a84b-4ea6-477e-88de-fceec3a95f31" containerName="collect-profiles" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.561709 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="c506a84b-4ea6-477e-88de-fceec3a95f31" containerName="collect-profiles" Oct 11 03:55:18 crc kubenswrapper[4967]: E1011 03:55:18.561732 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0f97fb5-caf9-44e8-b8f0-14e290c823e5" containerName="extract-content" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.561749 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0f97fb5-caf9-44e8-b8f0-14e290c823e5" containerName="extract-content" Oct 11 03:55:18 crc kubenswrapper[4967]: E1011 03:55:18.561770 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="68f026ee-3e09-46ae-b4af-a92ec0980e21" containerName="registry-server" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.561787 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="68f026ee-3e09-46ae-b4af-a92ec0980e21" containerName="registry-server" Oct 11 03:55:18 crc kubenswrapper[4967]: E1011 03:55:18.561809 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="572eea97-2918-420e-821b-b23b212191b3" containerName="extract-content" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.561825 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="572eea97-2918-420e-821b-b23b212191b3" containerName="extract-content" Oct 11 03:55:18 crc kubenswrapper[4967]: E1011 03:55:18.561855 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0dae89fc-445f-4531-8928-29b334636522" containerName="oauth-openshift" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.561873 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="0dae89fc-445f-4531-8928-29b334636522" containerName="oauth-openshift" Oct 11 03:55:18 crc kubenswrapper[4967]: E1011 03:55:18.561900 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="46038ef8-bdeb-4ff7-a5b4-bd09512ee26f" containerName="pruner" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.561916 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="46038ef8-bdeb-4ff7-a5b4-bd09512ee26f" containerName="pruner" Oct 11 03:55:18 crc kubenswrapper[4967]: E1011 03:55:18.561944 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e0f97fb5-caf9-44e8-b8f0-14e290c823e5" containerName="extract-utilities" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.561962 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="e0f97fb5-caf9-44e8-b8f0-14e290c823e5" containerName="extract-utilities" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.562224 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="5af71eaf-029a-4a8a-a563-e17c6eaeb6ce" containerName="registry-server" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.562253 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="46038ef8-bdeb-4ff7-a5b4-bd09512ee26f" containerName="pruner" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.562274 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="68f026ee-3e09-46ae-b4af-a92ec0980e21" containerName="registry-server" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.562298 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="572eea97-2918-420e-821b-b23b212191b3" containerName="registry-server" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.562322 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="57630488-3eda-4f83-93c6-7c9c5c07a109" containerName="pruner" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.562344 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="e0f97fb5-caf9-44e8-b8f0-14e290c823e5" containerName="registry-server" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.562366 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="c506a84b-4ea6-477e-88de-fceec3a95f31" containerName="collect-profiles" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.562392 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="0dae89fc-445f-4531-8928-29b334636522" containerName="oauth-openshift" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.563216 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.578905 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl"] Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.611953 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0dae89fc-445f-4531-8928-29b334636522-audit-policies\") pod \"0dae89fc-445f-4531-8928-29b334636522\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.612044 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xtvtk\" (UniqueName: \"kubernetes.io/projected/0dae89fc-445f-4531-8928-29b334636522-kube-api-access-xtvtk\") pod \"0dae89fc-445f-4531-8928-29b334636522\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.612114 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-trusted-ca-bundle\") pod \"0dae89fc-445f-4531-8928-29b334636522\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.612148 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-cliconfig\") pod \"0dae89fc-445f-4531-8928-29b334636522\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.612190 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-user-template-error\") pod \"0dae89fc-445f-4531-8928-29b334636522\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.612257 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-router-certs\") pod \"0dae89fc-445f-4531-8928-29b334636522\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.612291 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-session\") pod \"0dae89fc-445f-4531-8928-29b334636522\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.612322 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-serving-cert\") pod \"0dae89fc-445f-4531-8928-29b334636522\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.612367 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-user-idp-0-file-data\") pod \"0dae89fc-445f-4531-8928-29b334636522\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.612430 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0dae89fc-445f-4531-8928-29b334636522-audit-dir\") pod \"0dae89fc-445f-4531-8928-29b334636522\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.612472 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-user-template-provider-selection\") pod \"0dae89fc-445f-4531-8928-29b334636522\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.612508 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-service-ca\") pod \"0dae89fc-445f-4531-8928-29b334636522\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.612737 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-user-template-login\") pod \"0dae89fc-445f-4531-8928-29b334636522\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.612795 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-ocp-branding-template\") pod \"0dae89fc-445f-4531-8928-29b334636522\" (UID: \"0dae89fc-445f-4531-8928-29b334636522\") " Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.614940 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0dae89fc-445f-4531-8928-29b334636522-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "0dae89fc-445f-4531-8928-29b334636522" (UID: "0dae89fc-445f-4531-8928-29b334636522"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.616117 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "0dae89fc-445f-4531-8928-29b334636522" (UID: "0dae89fc-445f-4531-8928-29b334636522"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.616664 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "0dae89fc-445f-4531-8928-29b334636522" (UID: "0dae89fc-445f-4531-8928-29b334636522"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.616663 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "0dae89fc-445f-4531-8928-29b334636522" (UID: "0dae89fc-445f-4531-8928-29b334636522"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.616714 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/0dae89fc-445f-4531-8928-29b334636522-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "0dae89fc-445f-4531-8928-29b334636522" (UID: "0dae89fc-445f-4531-8928-29b334636522"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.620725 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "0dae89fc-445f-4531-8928-29b334636522" (UID: "0dae89fc-445f-4531-8928-29b334636522"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.621618 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "0dae89fc-445f-4531-8928-29b334636522" (UID: "0dae89fc-445f-4531-8928-29b334636522"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.622259 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0dae89fc-445f-4531-8928-29b334636522-kube-api-access-xtvtk" (OuterVolumeSpecName: "kube-api-access-xtvtk") pod "0dae89fc-445f-4531-8928-29b334636522" (UID: "0dae89fc-445f-4531-8928-29b334636522"). InnerVolumeSpecName "kube-api-access-xtvtk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.623283 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "0dae89fc-445f-4531-8928-29b334636522" (UID: "0dae89fc-445f-4531-8928-29b334636522"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.624245 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "0dae89fc-445f-4531-8928-29b334636522" (UID: "0dae89fc-445f-4531-8928-29b334636522"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.624557 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "0dae89fc-445f-4531-8928-29b334636522" (UID: "0dae89fc-445f-4531-8928-29b334636522"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.624751 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "0dae89fc-445f-4531-8928-29b334636522" (UID: "0dae89fc-445f-4531-8928-29b334636522"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.626936 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "0dae89fc-445f-4531-8928-29b334636522" (UID: "0dae89fc-445f-4531-8928-29b334636522"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.627164 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "0dae89fc-445f-4531-8928-29b334636522" (UID: "0dae89fc-445f-4531-8928-29b334636522"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.714754 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.714834 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.714878 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-system-router-certs\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.714956 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-system-session\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.715005 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkgrn\" (UniqueName: \"kubernetes.io/projected/477ac565-d655-47d3-a29e-35b2291d6484-kube-api-access-pkgrn\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.715116 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/477ac565-d655-47d3-a29e-35b2291d6484-audit-dir\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.715316 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.715430 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-user-template-login\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.715473 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.715511 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/477ac565-d655-47d3-a29e-35b2291d6484-audit-policies\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.715573 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-user-template-error\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.715614 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-system-serving-cert\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.715651 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-system-service-ca\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.715692 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-system-cliconfig\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.715869 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.715927 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.715952 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.715972 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.715996 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.716018 4967 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/0dae89fc-445f-4531-8928-29b334636522-audit-dir\") on node \"crc\" DevicePath \"\"" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.716043 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.716066 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.716188 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.716212 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.716232 4967 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/0dae89fc-445f-4531-8928-29b334636522-audit-policies\") on node \"crc\" DevicePath \"\"" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.716252 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xtvtk\" (UniqueName: \"kubernetes.io/projected/0dae89fc-445f-4531-8928-29b334636522-kube-api-access-xtvtk\") on node \"crc\" DevicePath \"\"" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.716273 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.716292 4967 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/0dae89fc-445f-4531-8928-29b334636522-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.817859 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkgrn\" (UniqueName: \"kubernetes.io/projected/477ac565-d655-47d3-a29e-35b2291d6484-kube-api-access-pkgrn\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.817937 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/477ac565-d655-47d3-a29e-35b2291d6484-audit-dir\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.817994 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.818065 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-user-template-login\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.818174 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.818199 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/477ac565-d655-47d3-a29e-35b2291d6484-audit-dir\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.818228 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/477ac565-d655-47d3-a29e-35b2291d6484-audit-policies\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.818442 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-user-template-error\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.818501 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-system-serving-cert\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.818563 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-system-service-ca\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.818623 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-system-cliconfig\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.818771 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.818824 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.818888 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-system-router-certs\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.818984 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-system-session\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.820418 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-system-cliconfig\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.820582 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/477ac565-d655-47d3-a29e-35b2291d6484-audit-policies\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.821458 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.821525 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-system-service-ca\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.824859 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.825010 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-system-session\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.825643 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-user-template-error\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.826143 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.826200 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-system-serving-cert\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.826491 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-user-template-login\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.828664 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-system-router-certs\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.829292 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/477ac565-d655-47d3-a29e-35b2291d6484-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.843123 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkgrn\" (UniqueName: \"kubernetes.io/projected/477ac565-d655-47d3-a29e-35b2291d6484-kube-api-access-pkgrn\") pod \"oauth-openshift-64f9fb64bf-s7fxl\" (UID: \"477ac565-d655-47d3-a29e-35b2291d6484\") " pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:18 crc kubenswrapper[4967]: I1011 03:55:18.890253 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:19 crc kubenswrapper[4967]: I1011 03:55:19.201204 4967 generic.go:334] "Generic (PLEG): container finished" podID="0dae89fc-445f-4531-8928-29b334636522" containerID="21a94896d91097d19aeb6b46354b56d5640a05428b63933176e0a1178bb5f651" exitCode=0 Oct 11 03:55:19 crc kubenswrapper[4967]: I1011 03:55:19.201278 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" Oct 11 03:55:19 crc kubenswrapper[4967]: I1011 03:55:19.201300 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" event={"ID":"0dae89fc-445f-4531-8928-29b334636522","Type":"ContainerDied","Data":"21a94896d91097d19aeb6b46354b56d5640a05428b63933176e0a1178bb5f651"} Oct 11 03:55:19 crc kubenswrapper[4967]: I1011 03:55:19.201413 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-6sjfd" event={"ID":"0dae89fc-445f-4531-8928-29b334636522","Type":"ContainerDied","Data":"81dc34607ed3916b6a34b0cad97035721d5397b349c54feaf67139e5c9de87e8"} Oct 11 03:55:19 crc kubenswrapper[4967]: I1011 03:55:19.201473 4967 scope.go:117] "RemoveContainer" containerID="21a94896d91097d19aeb6b46354b56d5640a05428b63933176e0a1178bb5f651" Oct 11 03:55:19 crc kubenswrapper[4967]: I1011 03:55:19.242786 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6sjfd"] Oct 11 03:55:19 crc kubenswrapper[4967]: I1011 03:55:19.243319 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-6sjfd"] Oct 11 03:55:19 crc kubenswrapper[4967]: I1011 03:55:19.243149 4967 scope.go:117] "RemoveContainer" containerID="21a94896d91097d19aeb6b46354b56d5640a05428b63933176e0a1178bb5f651" Oct 11 03:55:19 crc kubenswrapper[4967]: E1011 03:55:19.244358 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"21a94896d91097d19aeb6b46354b56d5640a05428b63933176e0a1178bb5f651\": container with ID starting with 21a94896d91097d19aeb6b46354b56d5640a05428b63933176e0a1178bb5f651 not found: ID does not exist" containerID="21a94896d91097d19aeb6b46354b56d5640a05428b63933176e0a1178bb5f651" Oct 11 03:55:19 crc kubenswrapper[4967]: I1011 03:55:19.244397 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"21a94896d91097d19aeb6b46354b56d5640a05428b63933176e0a1178bb5f651"} err="failed to get container status \"21a94896d91097d19aeb6b46354b56d5640a05428b63933176e0a1178bb5f651\": rpc error: code = NotFound desc = could not find container \"21a94896d91097d19aeb6b46354b56d5640a05428b63933176e0a1178bb5f651\": container with ID starting with 21a94896d91097d19aeb6b46354b56d5640a05428b63933176e0a1178bb5f651 not found: ID does not exist" Oct 11 03:55:19 crc kubenswrapper[4967]: I1011 03:55:19.347826 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl"] Oct 11 03:55:19 crc kubenswrapper[4967]: W1011 03:55:19.354247 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod477ac565_d655_47d3_a29e_35b2291d6484.slice/crio-5ee3ddddc1cc2094bfb19ffce8ed69db224512f15630e5bbc03eee51f3a6c76b WatchSource:0}: Error finding container 5ee3ddddc1cc2094bfb19ffce8ed69db224512f15630e5bbc03eee51f3a6c76b: Status 404 returned error can't find the container with id 5ee3ddddc1cc2094bfb19ffce8ed69db224512f15630e5bbc03eee51f3a6c76b Oct 11 03:55:20 crc kubenswrapper[4967]: I1011 03:55:20.210023 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" event={"ID":"477ac565-d655-47d3-a29e-35b2291d6484","Type":"ContainerStarted","Data":"31087591a4540cd634451bd5b648df81fd54a6163448ec30f4a345eb6eedc102"} Oct 11 03:55:20 crc kubenswrapper[4967]: I1011 03:55:20.210092 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" event={"ID":"477ac565-d655-47d3-a29e-35b2291d6484","Type":"ContainerStarted","Data":"5ee3ddddc1cc2094bfb19ffce8ed69db224512f15630e5bbc03eee51f3a6c76b"} Oct 11 03:55:20 crc kubenswrapper[4967]: I1011 03:55:20.211483 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:20 crc kubenswrapper[4967]: I1011 03:55:20.227321 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" Oct 11 03:55:20 crc kubenswrapper[4967]: I1011 03:55:20.235170 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-64f9fb64bf-s7fxl" podStartSLOduration=27.235149229 podStartE2EDuration="27.235149229s" podCreationTimestamp="2025-10-11 03:54:53 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:55:20.232675578 +0000 UTC m=+248.195884511" watchObservedRunningTime="2025-10-11 03:55:20.235149229 +0000 UTC m=+248.198358162" Oct 11 03:55:20 crc kubenswrapper[4967]: I1011 03:55:20.825976 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0dae89fc-445f-4531-8928-29b334636522" path="/var/lib/kubelet/pods/0dae89fc-445f-4531-8928-29b334636522/volumes" Oct 11 03:55:37 crc kubenswrapper[4967]: I1011 03:55:37.681166 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zbm5d"] Oct 11 03:55:37 crc kubenswrapper[4967]: I1011 03:55:37.682283 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-zbm5d" podUID="130a69d7-5216-4017-9146-1940a1d58210" containerName="registry-server" containerID="cri-o://c6da23a3cca166867af38c21c9c0a7ae952b046c5c4c942e8b9ac1952ad01af4" gracePeriod=30 Oct 11 03:55:37 crc kubenswrapper[4967]: I1011 03:55:37.690862 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-799qw"] Oct 11 03:55:37 crc kubenswrapper[4967]: I1011 03:55:37.691159 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-799qw" podUID="d05dc46c-e2ec-4c4b-bfed-dd509f93a5df" containerName="registry-server" containerID="cri-o://a44900551f8820045090c5a4323ab5056e638679919c4e2e2907861091d4e005" gracePeriod=30 Oct 11 03:55:37 crc kubenswrapper[4967]: I1011 03:55:37.700093 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-49rxs"] Oct 11 03:55:37 crc kubenswrapper[4967]: I1011 03:55:37.700340 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-49rxs" podUID="b9509019-37bf-4674-af71-c259393eab0b" containerName="marketplace-operator" containerID="cri-o://0537240f76748e255545f491ddc4f3f03411c1717cb8e8b271954b3d53028eff" gracePeriod=30 Oct 11 03:55:37 crc kubenswrapper[4967]: I1011 03:55:37.717757 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xjxhr"] Oct 11 03:55:37 crc kubenswrapper[4967]: I1011 03:55:37.717990 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-xjxhr" podUID="f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac" containerName="registry-server" containerID="cri-o://0f6ada5b1586532169620a1612c62c2adbf3a4d2e52d6c2f7a6abfc351b4d535" gracePeriod=30 Oct 11 03:55:37 crc kubenswrapper[4967]: I1011 03:55:37.730541 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sp7h8"] Oct 11 03:55:37 crc kubenswrapper[4967]: I1011 03:55:37.731401 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sp7h8" Oct 11 03:55:37 crc kubenswrapper[4967]: I1011 03:55:37.735417 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sp7h8"] Oct 11 03:55:37 crc kubenswrapper[4967]: I1011 03:55:37.738268 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vr9z8"] Oct 11 03:55:37 crc kubenswrapper[4967]: I1011 03:55:37.738514 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-vr9z8" podUID="f70cf6aa-6c00-4507-98c3-4f385cc57522" containerName="registry-server" containerID="cri-o://eeaf8e1eb112fbeb11b2e4302eba226b06d6e6f1980762e3d027caac5a150ce6" gracePeriod=30 Oct 11 03:55:37 crc kubenswrapper[4967]: I1011 03:55:37.881620 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8fpc\" (UniqueName: \"kubernetes.io/projected/020dc4f8-c49b-4447-8373-426bf747cace-kube-api-access-j8fpc\") pod \"marketplace-operator-79b997595-sp7h8\" (UID: \"020dc4f8-c49b-4447-8373-426bf747cace\") " pod="openshift-marketplace/marketplace-operator-79b997595-sp7h8" Oct 11 03:55:37 crc kubenswrapper[4967]: I1011 03:55:37.881701 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/020dc4f8-c49b-4447-8373-426bf747cace-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sp7h8\" (UID: \"020dc4f8-c49b-4447-8373-426bf747cace\") " pod="openshift-marketplace/marketplace-operator-79b997595-sp7h8" Oct 11 03:55:37 crc kubenswrapper[4967]: I1011 03:55:37.881726 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/020dc4f8-c49b-4447-8373-426bf747cace-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sp7h8\" (UID: \"020dc4f8-c49b-4447-8373-426bf747cace\") " pod="openshift-marketplace/marketplace-operator-79b997595-sp7h8" Oct 11 03:55:37 crc kubenswrapper[4967]: I1011 03:55:37.983166 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8fpc\" (UniqueName: \"kubernetes.io/projected/020dc4f8-c49b-4447-8373-426bf747cace-kube-api-access-j8fpc\") pod \"marketplace-operator-79b997595-sp7h8\" (UID: \"020dc4f8-c49b-4447-8373-426bf747cace\") " pod="openshift-marketplace/marketplace-operator-79b997595-sp7h8" Oct 11 03:55:37 crc kubenswrapper[4967]: I1011 03:55:37.983226 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/020dc4f8-c49b-4447-8373-426bf747cace-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sp7h8\" (UID: \"020dc4f8-c49b-4447-8373-426bf747cace\") " pod="openshift-marketplace/marketplace-operator-79b997595-sp7h8" Oct 11 03:55:37 crc kubenswrapper[4967]: I1011 03:55:37.983255 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/020dc4f8-c49b-4447-8373-426bf747cace-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sp7h8\" (UID: \"020dc4f8-c49b-4447-8373-426bf747cace\") " pod="openshift-marketplace/marketplace-operator-79b997595-sp7h8" Oct 11 03:55:37 crc kubenswrapper[4967]: I1011 03:55:37.984305 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/020dc4f8-c49b-4447-8373-426bf747cace-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-sp7h8\" (UID: \"020dc4f8-c49b-4447-8373-426bf747cace\") " pod="openshift-marketplace/marketplace-operator-79b997595-sp7h8" Oct 11 03:55:37 crc kubenswrapper[4967]: I1011 03:55:37.988630 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/020dc4f8-c49b-4447-8373-426bf747cace-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-sp7h8\" (UID: \"020dc4f8-c49b-4447-8373-426bf747cace\") " pod="openshift-marketplace/marketplace-operator-79b997595-sp7h8" Oct 11 03:55:38 crc kubenswrapper[4967]: I1011 03:55:38.007003 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8fpc\" (UniqueName: \"kubernetes.io/projected/020dc4f8-c49b-4447-8373-426bf747cace-kube-api-access-j8fpc\") pod \"marketplace-operator-79b997595-sp7h8\" (UID: \"020dc4f8-c49b-4447-8373-426bf747cace\") " pod="openshift-marketplace/marketplace-operator-79b997595-sp7h8" Oct 11 03:55:38 crc kubenswrapper[4967]: I1011 03:55:38.069052 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-sp7h8" Oct 11 03:55:38 crc kubenswrapper[4967]: I1011 03:55:38.329890 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-sp7h8"] Oct 11 03:55:38 crc kubenswrapper[4967]: I1011 03:55:38.348323 4967 generic.go:334] "Generic (PLEG): container finished" podID="b9509019-37bf-4674-af71-c259393eab0b" containerID="0537240f76748e255545f491ddc4f3f03411c1717cb8e8b271954b3d53028eff" exitCode=0 Oct 11 03:55:38 crc kubenswrapper[4967]: I1011 03:55:38.348416 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-49rxs" event={"ID":"b9509019-37bf-4674-af71-c259393eab0b","Type":"ContainerDied","Data":"0537240f76748e255545f491ddc4f3f03411c1717cb8e8b271954b3d53028eff"} Oct 11 03:55:38 crc kubenswrapper[4967]: I1011 03:55:38.350945 4967 generic.go:334] "Generic (PLEG): container finished" podID="f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac" containerID="0f6ada5b1586532169620a1612c62c2adbf3a4d2e52d6c2f7a6abfc351b4d535" exitCode=0 Oct 11 03:55:38 crc kubenswrapper[4967]: I1011 03:55:38.351039 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xjxhr" event={"ID":"f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac","Type":"ContainerDied","Data":"0f6ada5b1586532169620a1612c62c2adbf3a4d2e52d6c2f7a6abfc351b4d535"} Oct 11 03:55:38 crc kubenswrapper[4967]: I1011 03:55:38.352871 4967 generic.go:334] "Generic (PLEG): container finished" podID="d05dc46c-e2ec-4c4b-bfed-dd509f93a5df" containerID="a44900551f8820045090c5a4323ab5056e638679919c4e2e2907861091d4e005" exitCode=0 Oct 11 03:55:38 crc kubenswrapper[4967]: I1011 03:55:38.352884 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-799qw" event={"ID":"d05dc46c-e2ec-4c4b-bfed-dd509f93a5df","Type":"ContainerDied","Data":"a44900551f8820045090c5a4323ab5056e638679919c4e2e2907861091d4e005"} Oct 11 03:55:38 crc kubenswrapper[4967]: I1011 03:55:38.354773 4967 generic.go:334] "Generic (PLEG): container finished" podID="130a69d7-5216-4017-9146-1940a1d58210" containerID="c6da23a3cca166867af38c21c9c0a7ae952b046c5c4c942e8b9ac1952ad01af4" exitCode=0 Oct 11 03:55:38 crc kubenswrapper[4967]: I1011 03:55:38.354803 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zbm5d" event={"ID":"130a69d7-5216-4017-9146-1940a1d58210","Type":"ContainerDied","Data":"c6da23a3cca166867af38c21c9c0a7ae952b046c5c4c942e8b9ac1952ad01af4"} Oct 11 03:55:38 crc kubenswrapper[4967]: W1011 03:55:38.413849 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod020dc4f8_c49b_4447_8373_426bf747cace.slice/crio-26f09b9dd758559a6857b80a65cd349f2380078bb6e656c4be0d52b6c0687a5a WatchSource:0}: Error finding container 26f09b9dd758559a6857b80a65cd349f2380078bb6e656c4be0d52b6c0687a5a: Status 404 returned error can't find the container with id 26f09b9dd758559a6857b80a65cd349f2380078bb6e656c4be0d52b6c0687a5a Oct 11 03:55:38 crc kubenswrapper[4967]: I1011 03:55:38.641065 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-799qw" Oct 11 03:55:38 crc kubenswrapper[4967]: I1011 03:55:38.697552 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xjxhr" Oct 11 03:55:38 crc kubenswrapper[4967]: I1011 03:55:38.793556 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d05dc46c-e2ec-4c4b-bfed-dd509f93a5df-utilities\") pod \"d05dc46c-e2ec-4c4b-bfed-dd509f93a5df\" (UID: \"d05dc46c-e2ec-4c4b-bfed-dd509f93a5df\") " Oct 11 03:55:38 crc kubenswrapper[4967]: I1011 03:55:38.793626 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d05dc46c-e2ec-4c4b-bfed-dd509f93a5df-catalog-content\") pod \"d05dc46c-e2ec-4c4b-bfed-dd509f93a5df\" (UID: \"d05dc46c-e2ec-4c4b-bfed-dd509f93a5df\") " Oct 11 03:55:38 crc kubenswrapper[4967]: I1011 03:55:38.793658 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fx4sl\" (UniqueName: \"kubernetes.io/projected/d05dc46c-e2ec-4c4b-bfed-dd509f93a5df-kube-api-access-fx4sl\") pod \"d05dc46c-e2ec-4c4b-bfed-dd509f93a5df\" (UID: \"d05dc46c-e2ec-4c4b-bfed-dd509f93a5df\") " Oct 11 03:55:38 crc kubenswrapper[4967]: I1011 03:55:38.794936 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d05dc46c-e2ec-4c4b-bfed-dd509f93a5df-utilities" (OuterVolumeSpecName: "utilities") pod "d05dc46c-e2ec-4c4b-bfed-dd509f93a5df" (UID: "d05dc46c-e2ec-4c4b-bfed-dd509f93a5df"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:55:38 crc kubenswrapper[4967]: I1011 03:55:38.799318 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d05dc46c-e2ec-4c4b-bfed-dd509f93a5df-kube-api-access-fx4sl" (OuterVolumeSpecName: "kube-api-access-fx4sl") pod "d05dc46c-e2ec-4c4b-bfed-dd509f93a5df" (UID: "d05dc46c-e2ec-4c4b-bfed-dd509f93a5df"). InnerVolumeSpecName "kube-api-access-fx4sl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:55:38 crc kubenswrapper[4967]: I1011 03:55:38.853799 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d05dc46c-e2ec-4c4b-bfed-dd509f93a5df-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "d05dc46c-e2ec-4c4b-bfed-dd509f93a5df" (UID: "d05dc46c-e2ec-4c4b-bfed-dd509f93a5df"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:55:38 crc kubenswrapper[4967]: I1011 03:55:38.894975 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac-catalog-content\") pod \"f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac\" (UID: \"f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac\") " Oct 11 03:55:38 crc kubenswrapper[4967]: I1011 03:55:38.895146 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lws4d\" (UniqueName: \"kubernetes.io/projected/f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac-kube-api-access-lws4d\") pod \"f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac\" (UID: \"f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac\") " Oct 11 03:55:38 crc kubenswrapper[4967]: I1011 03:55:38.895178 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac-utilities\") pod \"f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac\" (UID: \"f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac\") " Oct 11 03:55:38 crc kubenswrapper[4967]: I1011 03:55:38.895369 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/d05dc46c-e2ec-4c4b-bfed-dd509f93a5df-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:55:38 crc kubenswrapper[4967]: I1011 03:55:38.895385 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/d05dc46c-e2ec-4c4b-bfed-dd509f93a5df-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:55:38 crc kubenswrapper[4967]: I1011 03:55:38.895399 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fx4sl\" (UniqueName: \"kubernetes.io/projected/d05dc46c-e2ec-4c4b-bfed-dd509f93a5df-kube-api-access-fx4sl\") on node \"crc\" DevicePath \"\"" Oct 11 03:55:38 crc kubenswrapper[4967]: I1011 03:55:38.896128 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac-utilities" (OuterVolumeSpecName: "utilities") pod "f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac" (UID: "f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:55:38 crc kubenswrapper[4967]: I1011 03:55:38.898114 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac-kube-api-access-lws4d" (OuterVolumeSpecName: "kube-api-access-lws4d") pod "f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac" (UID: "f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac"). InnerVolumeSpecName "kube-api-access-lws4d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:55:38 crc kubenswrapper[4967]: I1011 03:55:38.913106 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac" (UID: "f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:55:38 crc kubenswrapper[4967]: I1011 03:55:38.996557 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lws4d\" (UniqueName: \"kubernetes.io/projected/f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac-kube-api-access-lws4d\") on node \"crc\" DevicePath \"\"" Oct 11 03:55:38 crc kubenswrapper[4967]: I1011 03:55:38.996594 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:55:38 crc kubenswrapper[4967]: I1011 03:55:38.996604 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.167859 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zbm5d" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.271644 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-49rxs" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.276279 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vr9z8" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.299270 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/130a69d7-5216-4017-9146-1940a1d58210-utilities\") pod \"130a69d7-5216-4017-9146-1940a1d58210\" (UID: \"130a69d7-5216-4017-9146-1940a1d58210\") " Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.299345 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vsqz9\" (UniqueName: \"kubernetes.io/projected/130a69d7-5216-4017-9146-1940a1d58210-kube-api-access-vsqz9\") pod \"130a69d7-5216-4017-9146-1940a1d58210\" (UID: \"130a69d7-5216-4017-9146-1940a1d58210\") " Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.299449 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/130a69d7-5216-4017-9146-1940a1d58210-catalog-content\") pod \"130a69d7-5216-4017-9146-1940a1d58210\" (UID: \"130a69d7-5216-4017-9146-1940a1d58210\") " Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.300156 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/130a69d7-5216-4017-9146-1940a1d58210-utilities" (OuterVolumeSpecName: "utilities") pod "130a69d7-5216-4017-9146-1940a1d58210" (UID: "130a69d7-5216-4017-9146-1940a1d58210"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.308475 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/130a69d7-5216-4017-9146-1940a1d58210-kube-api-access-vsqz9" (OuterVolumeSpecName: "kube-api-access-vsqz9") pod "130a69d7-5216-4017-9146-1940a1d58210" (UID: "130a69d7-5216-4017-9146-1940a1d58210"). InnerVolumeSpecName "kube-api-access-vsqz9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.356884 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/130a69d7-5216-4017-9146-1940a1d58210-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "130a69d7-5216-4017-9146-1940a1d58210" (UID: "130a69d7-5216-4017-9146-1940a1d58210"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.363282 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-49rxs" event={"ID":"b9509019-37bf-4674-af71-c259393eab0b","Type":"ContainerDied","Data":"09dd985e0fa0ae2f2cbb5ea8523da21f78bc2436ad959c2f3bf1d8b812b2e01d"} Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.363336 4967 scope.go:117] "RemoveContainer" containerID="0537240f76748e255545f491ddc4f3f03411c1717cb8e8b271954b3d53028eff" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.363447 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-49rxs" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.370310 4967 generic.go:334] "Generic (PLEG): container finished" podID="f70cf6aa-6c00-4507-98c3-4f385cc57522" containerID="eeaf8e1eb112fbeb11b2e4302eba226b06d6e6f1980762e3d027caac5a150ce6" exitCode=0 Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.370446 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vr9z8" event={"ID":"f70cf6aa-6c00-4507-98c3-4f385cc57522","Type":"ContainerDied","Data":"eeaf8e1eb112fbeb11b2e4302eba226b06d6e6f1980762e3d027caac5a150ce6"} Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.370480 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vr9z8" event={"ID":"f70cf6aa-6c00-4507-98c3-4f385cc57522","Type":"ContainerDied","Data":"bab94f0f0c3b481e24f11c67cc7e2a4a657321e134eec71be13d3d99bd92dc43"} Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.370620 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vr9z8" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.375394 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-xjxhr" event={"ID":"f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac","Type":"ContainerDied","Data":"855379c72f6d636d494e86e891bca308bddf0eb428a87fd239f2ec8b8afa528f"} Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.375525 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-xjxhr" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.380400 4967 scope.go:117] "RemoveContainer" containerID="eeaf8e1eb112fbeb11b2e4302eba226b06d6e6f1980762e3d027caac5a150ce6" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.381838 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-799qw" event={"ID":"d05dc46c-e2ec-4c4b-bfed-dd509f93a5df","Type":"ContainerDied","Data":"5c224038aef95047b2222fca8bd4225e68df6b467265ec804bb5c595b0947c38"} Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.381908 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-799qw" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.389539 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zbm5d" event={"ID":"130a69d7-5216-4017-9146-1940a1d58210","Type":"ContainerDied","Data":"bad7f074e36a83b3f576de8abd25928f38f125d938fcc2613a2fcb75936b8110"} Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.389640 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zbm5d" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.400537 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f70cf6aa-6c00-4507-98c3-4f385cc57522-catalog-content\") pod \"f70cf6aa-6c00-4507-98c3-4f385cc57522\" (UID: \"f70cf6aa-6c00-4507-98c3-4f385cc57522\") " Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.400614 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f70cf6aa-6c00-4507-98c3-4f385cc57522-utilities\") pod \"f70cf6aa-6c00-4507-98c3-4f385cc57522\" (UID: \"f70cf6aa-6c00-4507-98c3-4f385cc57522\") " Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.400668 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b9509019-37bf-4674-af71-c259393eab0b-marketplace-trusted-ca\") pod \"b9509019-37bf-4674-af71-c259393eab0b\" (UID: \"b9509019-37bf-4674-af71-c259393eab0b\") " Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.400702 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b9509019-37bf-4674-af71-c259393eab0b-marketplace-operator-metrics\") pod \"b9509019-37bf-4674-af71-c259393eab0b\" (UID: \"b9509019-37bf-4674-af71-c259393eab0b\") " Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.400743 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tj978\" (UniqueName: \"kubernetes.io/projected/f70cf6aa-6c00-4507-98c3-4f385cc57522-kube-api-access-tj978\") pod \"f70cf6aa-6c00-4507-98c3-4f385cc57522\" (UID: \"f70cf6aa-6c00-4507-98c3-4f385cc57522\") " Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.400793 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-skjkj\" (UniqueName: \"kubernetes.io/projected/b9509019-37bf-4674-af71-c259393eab0b-kube-api-access-skjkj\") pod \"b9509019-37bf-4674-af71-c259393eab0b\" (UID: \"b9509019-37bf-4674-af71-c259393eab0b\") " Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.401046 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/130a69d7-5216-4017-9146-1940a1d58210-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.401076 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/130a69d7-5216-4017-9146-1940a1d58210-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.401147 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vsqz9\" (UniqueName: \"kubernetes.io/projected/130a69d7-5216-4017-9146-1940a1d58210-kube-api-access-vsqz9\") on node \"crc\" DevicePath \"\"" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.401480 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9509019-37bf-4674-af71-c259393eab0b-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b9509019-37bf-4674-af71-c259393eab0b" (UID: "b9509019-37bf-4674-af71-c259393eab0b"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.401708 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f70cf6aa-6c00-4507-98c3-4f385cc57522-utilities" (OuterVolumeSpecName: "utilities") pod "f70cf6aa-6c00-4507-98c3-4f385cc57522" (UID: "f70cf6aa-6c00-4507-98c3-4f385cc57522"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.404677 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9509019-37bf-4674-af71-c259393eab0b-kube-api-access-skjkj" (OuterVolumeSpecName: "kube-api-access-skjkj") pod "b9509019-37bf-4674-af71-c259393eab0b" (UID: "b9509019-37bf-4674-af71-c259393eab0b"). InnerVolumeSpecName "kube-api-access-skjkj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.407645 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f70cf6aa-6c00-4507-98c3-4f385cc57522-kube-api-access-tj978" (OuterVolumeSpecName: "kube-api-access-tj978") pod "f70cf6aa-6c00-4507-98c3-4f385cc57522" (UID: "f70cf6aa-6c00-4507-98c3-4f385cc57522"). InnerVolumeSpecName "kube-api-access-tj978". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.416565 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sp7h8" event={"ID":"020dc4f8-c49b-4447-8373-426bf747cace","Type":"ContainerStarted","Data":"d1182b49fdc001342884c73a20427372bcb20743a34f72893bed80482ffddb09"} Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.416611 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-sp7h8" event={"ID":"020dc4f8-c49b-4447-8373-426bf747cace","Type":"ContainerStarted","Data":"26f09b9dd758559a6857b80a65cd349f2380078bb6e656c4be0d52b6c0687a5a"} Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.417481 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-sp7h8" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.426487 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-sp7h8" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.442740 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b9509019-37bf-4674-af71-c259393eab0b-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b9509019-37bf-4674-af71-c259393eab0b" (UID: "b9509019-37bf-4674-af71-c259393eab0b"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.443018 4967 scope.go:117] "RemoveContainer" containerID="9eaf587a3d43f9a38d7ddb02adadc6b95b93dfd979b538d7908d39d25729c31e" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.443334 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-799qw"] Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.472270 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-799qw"] Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.501967 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f70cf6aa-6c00-4507-98c3-4f385cc57522-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.502022 4967 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b9509019-37bf-4674-af71-c259393eab0b-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.502036 4967 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b9509019-37bf-4674-af71-c259393eab0b-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.502048 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tj978\" (UniqueName: \"kubernetes.io/projected/f70cf6aa-6c00-4507-98c3-4f385cc57522-kube-api-access-tj978\") on node \"crc\" DevicePath \"\"" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.502060 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-skjkj\" (UniqueName: \"kubernetes.io/projected/b9509019-37bf-4674-af71-c259393eab0b-kube-api-access-skjkj\") on node \"crc\" DevicePath \"\"" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.502958 4967 scope.go:117] "RemoveContainer" containerID="f6d5363b33e25ec5859961d37df45e033ef6e51d22836921080fe09b32dabd13" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.507187 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-xjxhr"] Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.521166 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-xjxhr"] Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.522566 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-sp7h8" podStartSLOduration=2.522552637 podStartE2EDuration="2.522552637s" podCreationTimestamp="2025-10-11 03:55:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:55:39.485559111 +0000 UTC m=+267.448768044" watchObservedRunningTime="2025-10-11 03:55:39.522552637 +0000 UTC m=+267.485761570" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.531883 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f70cf6aa-6c00-4507-98c3-4f385cc57522-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f70cf6aa-6c00-4507-98c3-4f385cc57522" (UID: "f70cf6aa-6c00-4507-98c3-4f385cc57522"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.545944 4967 scope.go:117] "RemoveContainer" containerID="eeaf8e1eb112fbeb11b2e4302eba226b06d6e6f1980762e3d027caac5a150ce6" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.550259 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-zbm5d"] Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.550523 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-zbm5d"] Oct 11 03:55:39 crc kubenswrapper[4967]: E1011 03:55:39.550643 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eeaf8e1eb112fbeb11b2e4302eba226b06d6e6f1980762e3d027caac5a150ce6\": container with ID starting with eeaf8e1eb112fbeb11b2e4302eba226b06d6e6f1980762e3d027caac5a150ce6 not found: ID does not exist" containerID="eeaf8e1eb112fbeb11b2e4302eba226b06d6e6f1980762e3d027caac5a150ce6" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.550724 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eeaf8e1eb112fbeb11b2e4302eba226b06d6e6f1980762e3d027caac5a150ce6"} err="failed to get container status \"eeaf8e1eb112fbeb11b2e4302eba226b06d6e6f1980762e3d027caac5a150ce6\": rpc error: code = NotFound desc = could not find container \"eeaf8e1eb112fbeb11b2e4302eba226b06d6e6f1980762e3d027caac5a150ce6\": container with ID starting with eeaf8e1eb112fbeb11b2e4302eba226b06d6e6f1980762e3d027caac5a150ce6 not found: ID does not exist" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.550811 4967 scope.go:117] "RemoveContainer" containerID="9eaf587a3d43f9a38d7ddb02adadc6b95b93dfd979b538d7908d39d25729c31e" Oct 11 03:55:39 crc kubenswrapper[4967]: E1011 03:55:39.554255 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9eaf587a3d43f9a38d7ddb02adadc6b95b93dfd979b538d7908d39d25729c31e\": container with ID starting with 9eaf587a3d43f9a38d7ddb02adadc6b95b93dfd979b538d7908d39d25729c31e not found: ID does not exist" containerID="9eaf587a3d43f9a38d7ddb02adadc6b95b93dfd979b538d7908d39d25729c31e" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.554295 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9eaf587a3d43f9a38d7ddb02adadc6b95b93dfd979b538d7908d39d25729c31e"} err="failed to get container status \"9eaf587a3d43f9a38d7ddb02adadc6b95b93dfd979b538d7908d39d25729c31e\": rpc error: code = NotFound desc = could not find container \"9eaf587a3d43f9a38d7ddb02adadc6b95b93dfd979b538d7908d39d25729c31e\": container with ID starting with 9eaf587a3d43f9a38d7ddb02adadc6b95b93dfd979b538d7908d39d25729c31e not found: ID does not exist" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.554321 4967 scope.go:117] "RemoveContainer" containerID="f6d5363b33e25ec5859961d37df45e033ef6e51d22836921080fe09b32dabd13" Oct 11 03:55:39 crc kubenswrapper[4967]: E1011 03:55:39.555390 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6d5363b33e25ec5859961d37df45e033ef6e51d22836921080fe09b32dabd13\": container with ID starting with f6d5363b33e25ec5859961d37df45e033ef6e51d22836921080fe09b32dabd13 not found: ID does not exist" containerID="f6d5363b33e25ec5859961d37df45e033ef6e51d22836921080fe09b32dabd13" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.555412 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6d5363b33e25ec5859961d37df45e033ef6e51d22836921080fe09b32dabd13"} err="failed to get container status \"f6d5363b33e25ec5859961d37df45e033ef6e51d22836921080fe09b32dabd13\": rpc error: code = NotFound desc = could not find container \"f6d5363b33e25ec5859961d37df45e033ef6e51d22836921080fe09b32dabd13\": container with ID starting with f6d5363b33e25ec5859961d37df45e033ef6e51d22836921080fe09b32dabd13 not found: ID does not exist" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.555430 4967 scope.go:117] "RemoveContainer" containerID="0f6ada5b1586532169620a1612c62c2adbf3a4d2e52d6c2f7a6abfc351b4d535" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.567814 4967 scope.go:117] "RemoveContainer" containerID="4b7afad0f405055b4293e3c4950d487dfe0f0ba03ed20ebf9f7e4fc8041552de" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.589467 4967 scope.go:117] "RemoveContainer" containerID="10943eaca91a949c0d167198c27ed265fa1318ce4219cab5bd17308a8aa32577" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.603005 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f70cf6aa-6c00-4507-98c3-4f385cc57522-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.608323 4967 scope.go:117] "RemoveContainer" containerID="a44900551f8820045090c5a4323ab5056e638679919c4e2e2907861091d4e005" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.619934 4967 scope.go:117] "RemoveContainer" containerID="7961c1c0282827e842a607993d2202e88cc54fbad70c26ebe17d5e0485ec6625" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.640231 4967 scope.go:117] "RemoveContainer" containerID="00c6be927f7455bc22fdd520a0c4f146bbe7bb1cc95debb3637831fdb5ea1f13" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.656111 4967 scope.go:117] "RemoveContainer" containerID="c6da23a3cca166867af38c21c9c0a7ae952b046c5c4c942e8b9ac1952ad01af4" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.668477 4967 scope.go:117] "RemoveContainer" containerID="85a60964d4075d899e009d9edf9be19ba25286ca3247114709d8e187293aae6e" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.683510 4967 scope.go:117] "RemoveContainer" containerID="6bb2124bf15e4b330f16e40c088c5045b674c6c278cb382497c812d7d5281f1d" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.691573 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-49rxs"] Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.696721 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-49rxs"] Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.708323 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-vr9z8"] Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.713586 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-vr9z8"] Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.896928 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-p2xpx"] Oct 11 03:55:39 crc kubenswrapper[4967]: E1011 03:55:39.897432 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac" containerName="extract-content" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.897569 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac" containerName="extract-content" Oct 11 03:55:39 crc kubenswrapper[4967]: E1011 03:55:39.897660 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f70cf6aa-6c00-4507-98c3-4f385cc57522" containerName="extract-utilities" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.897740 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f70cf6aa-6c00-4507-98c3-4f385cc57522" containerName="extract-utilities" Oct 11 03:55:39 crc kubenswrapper[4967]: E1011 03:55:39.897813 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d05dc46c-e2ec-4c4b-bfed-dd509f93a5df" containerName="registry-server" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.897887 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="d05dc46c-e2ec-4c4b-bfed-dd509f93a5df" containerName="registry-server" Oct 11 03:55:39 crc kubenswrapper[4967]: E1011 03:55:39.897964 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac" containerName="extract-utilities" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.898033 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac" containerName="extract-utilities" Oct 11 03:55:39 crc kubenswrapper[4967]: E1011 03:55:39.898124 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="130a69d7-5216-4017-9146-1940a1d58210" containerName="extract-utilities" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.898226 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="130a69d7-5216-4017-9146-1940a1d58210" containerName="extract-utilities" Oct 11 03:55:39 crc kubenswrapper[4967]: E1011 03:55:39.898301 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d05dc46c-e2ec-4c4b-bfed-dd509f93a5df" containerName="extract-utilities" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.898371 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="d05dc46c-e2ec-4c4b-bfed-dd509f93a5df" containerName="extract-utilities" Oct 11 03:55:39 crc kubenswrapper[4967]: E1011 03:55:39.898455 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac" containerName="registry-server" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.898527 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac" containerName="registry-server" Oct 11 03:55:39 crc kubenswrapper[4967]: E1011 03:55:39.898606 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f70cf6aa-6c00-4507-98c3-4f385cc57522" containerName="extract-content" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.898675 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f70cf6aa-6c00-4507-98c3-4f385cc57522" containerName="extract-content" Oct 11 03:55:39 crc kubenswrapper[4967]: E1011 03:55:39.898751 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="130a69d7-5216-4017-9146-1940a1d58210" containerName="extract-content" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.898826 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="130a69d7-5216-4017-9146-1940a1d58210" containerName="extract-content" Oct 11 03:55:39 crc kubenswrapper[4967]: E1011 03:55:39.898900 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9509019-37bf-4674-af71-c259393eab0b" containerName="marketplace-operator" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.898975 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9509019-37bf-4674-af71-c259393eab0b" containerName="marketplace-operator" Oct 11 03:55:39 crc kubenswrapper[4967]: E1011 03:55:39.899050 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="130a69d7-5216-4017-9146-1940a1d58210" containerName="registry-server" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.899159 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="130a69d7-5216-4017-9146-1940a1d58210" containerName="registry-server" Oct 11 03:55:39 crc kubenswrapper[4967]: E1011 03:55:39.899240 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f70cf6aa-6c00-4507-98c3-4f385cc57522" containerName="registry-server" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.899315 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f70cf6aa-6c00-4507-98c3-4f385cc57522" containerName="registry-server" Oct 11 03:55:39 crc kubenswrapper[4967]: E1011 03:55:39.899393 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d05dc46c-e2ec-4c4b-bfed-dd509f93a5df" containerName="extract-content" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.899468 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="d05dc46c-e2ec-4c4b-bfed-dd509f93a5df" containerName="extract-content" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.899640 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="130a69d7-5216-4017-9146-1940a1d58210" containerName="registry-server" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.899726 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9509019-37bf-4674-af71-c259393eab0b" containerName="marketplace-operator" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.899799 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac" containerName="registry-server" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.899871 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="f70cf6aa-6c00-4507-98c3-4f385cc57522" containerName="registry-server" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.899944 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="d05dc46c-e2ec-4c4b-bfed-dd509f93a5df" containerName="registry-server" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.900853 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p2xpx" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.903353 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Oct 11 03:55:39 crc kubenswrapper[4967]: I1011 03:55:39.905325 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p2xpx"] Oct 11 03:55:40 crc kubenswrapper[4967]: I1011 03:55:40.015302 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vt2lm\" (UniqueName: \"kubernetes.io/projected/065a2523-459f-440e-b5da-5ea5318fd89f-kube-api-access-vt2lm\") pod \"redhat-marketplace-p2xpx\" (UID: \"065a2523-459f-440e-b5da-5ea5318fd89f\") " pod="openshift-marketplace/redhat-marketplace-p2xpx" Oct 11 03:55:40 crc kubenswrapper[4967]: I1011 03:55:40.015369 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/065a2523-459f-440e-b5da-5ea5318fd89f-catalog-content\") pod \"redhat-marketplace-p2xpx\" (UID: \"065a2523-459f-440e-b5da-5ea5318fd89f\") " pod="openshift-marketplace/redhat-marketplace-p2xpx" Oct 11 03:55:40 crc kubenswrapper[4967]: I1011 03:55:40.015428 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/065a2523-459f-440e-b5da-5ea5318fd89f-utilities\") pod \"redhat-marketplace-p2xpx\" (UID: \"065a2523-459f-440e-b5da-5ea5318fd89f\") " pod="openshift-marketplace/redhat-marketplace-p2xpx" Oct 11 03:55:40 crc kubenswrapper[4967]: I1011 03:55:40.099807 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mxkbg"] Oct 11 03:55:40 crc kubenswrapper[4967]: I1011 03:55:40.101289 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mxkbg" Oct 11 03:55:40 crc kubenswrapper[4967]: I1011 03:55:40.104045 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Oct 11 03:55:40 crc kubenswrapper[4967]: I1011 03:55:40.110554 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mxkbg"] Oct 11 03:55:40 crc kubenswrapper[4967]: I1011 03:55:40.156948 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-db2px\" (UniqueName: \"kubernetes.io/projected/afb60be7-f83c-4dc3-9129-ec66a8b75a55-kube-api-access-db2px\") pod \"certified-operators-mxkbg\" (UID: \"afb60be7-f83c-4dc3-9129-ec66a8b75a55\") " pod="openshift-marketplace/certified-operators-mxkbg" Oct 11 03:55:40 crc kubenswrapper[4967]: I1011 03:55:40.157055 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afb60be7-f83c-4dc3-9129-ec66a8b75a55-utilities\") pod \"certified-operators-mxkbg\" (UID: \"afb60be7-f83c-4dc3-9129-ec66a8b75a55\") " pod="openshift-marketplace/certified-operators-mxkbg" Oct 11 03:55:40 crc kubenswrapper[4967]: I1011 03:55:40.157206 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vt2lm\" (UniqueName: \"kubernetes.io/projected/065a2523-459f-440e-b5da-5ea5318fd89f-kube-api-access-vt2lm\") pod \"redhat-marketplace-p2xpx\" (UID: \"065a2523-459f-440e-b5da-5ea5318fd89f\") " pod="openshift-marketplace/redhat-marketplace-p2xpx" Oct 11 03:55:40 crc kubenswrapper[4967]: I1011 03:55:40.157334 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/065a2523-459f-440e-b5da-5ea5318fd89f-catalog-content\") pod \"redhat-marketplace-p2xpx\" (UID: \"065a2523-459f-440e-b5da-5ea5318fd89f\") " pod="openshift-marketplace/redhat-marketplace-p2xpx" Oct 11 03:55:40 crc kubenswrapper[4967]: I1011 03:55:40.157481 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afb60be7-f83c-4dc3-9129-ec66a8b75a55-catalog-content\") pod \"certified-operators-mxkbg\" (UID: \"afb60be7-f83c-4dc3-9129-ec66a8b75a55\") " pod="openshift-marketplace/certified-operators-mxkbg" Oct 11 03:55:40 crc kubenswrapper[4967]: I1011 03:55:40.157960 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/065a2523-459f-440e-b5da-5ea5318fd89f-catalog-content\") pod \"redhat-marketplace-p2xpx\" (UID: \"065a2523-459f-440e-b5da-5ea5318fd89f\") " pod="openshift-marketplace/redhat-marketplace-p2xpx" Oct 11 03:55:40 crc kubenswrapper[4967]: I1011 03:55:40.159192 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/065a2523-459f-440e-b5da-5ea5318fd89f-utilities\") pod \"redhat-marketplace-p2xpx\" (UID: \"065a2523-459f-440e-b5da-5ea5318fd89f\") " pod="openshift-marketplace/redhat-marketplace-p2xpx" Oct 11 03:55:40 crc kubenswrapper[4967]: I1011 03:55:40.159594 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/065a2523-459f-440e-b5da-5ea5318fd89f-utilities\") pod \"redhat-marketplace-p2xpx\" (UID: \"065a2523-459f-440e-b5da-5ea5318fd89f\") " pod="openshift-marketplace/redhat-marketplace-p2xpx" Oct 11 03:55:40 crc kubenswrapper[4967]: I1011 03:55:40.184118 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vt2lm\" (UniqueName: \"kubernetes.io/projected/065a2523-459f-440e-b5da-5ea5318fd89f-kube-api-access-vt2lm\") pod \"redhat-marketplace-p2xpx\" (UID: \"065a2523-459f-440e-b5da-5ea5318fd89f\") " pod="openshift-marketplace/redhat-marketplace-p2xpx" Oct 11 03:55:40 crc kubenswrapper[4967]: I1011 03:55:40.219032 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-p2xpx" Oct 11 03:55:40 crc kubenswrapper[4967]: I1011 03:55:40.261499 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afb60be7-f83c-4dc3-9129-ec66a8b75a55-catalog-content\") pod \"certified-operators-mxkbg\" (UID: \"afb60be7-f83c-4dc3-9129-ec66a8b75a55\") " pod="openshift-marketplace/certified-operators-mxkbg" Oct 11 03:55:40 crc kubenswrapper[4967]: I1011 03:55:40.261606 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-db2px\" (UniqueName: \"kubernetes.io/projected/afb60be7-f83c-4dc3-9129-ec66a8b75a55-kube-api-access-db2px\") pod \"certified-operators-mxkbg\" (UID: \"afb60be7-f83c-4dc3-9129-ec66a8b75a55\") " pod="openshift-marketplace/certified-operators-mxkbg" Oct 11 03:55:40 crc kubenswrapper[4967]: I1011 03:55:40.263249 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afb60be7-f83c-4dc3-9129-ec66a8b75a55-utilities\") pod \"certified-operators-mxkbg\" (UID: \"afb60be7-f83c-4dc3-9129-ec66a8b75a55\") " pod="openshift-marketplace/certified-operators-mxkbg" Oct 11 03:55:40 crc kubenswrapper[4967]: I1011 03:55:40.263950 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/afb60be7-f83c-4dc3-9129-ec66a8b75a55-utilities\") pod \"certified-operators-mxkbg\" (UID: \"afb60be7-f83c-4dc3-9129-ec66a8b75a55\") " pod="openshift-marketplace/certified-operators-mxkbg" Oct 11 03:55:40 crc kubenswrapper[4967]: I1011 03:55:40.264489 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/afb60be7-f83c-4dc3-9129-ec66a8b75a55-catalog-content\") pod \"certified-operators-mxkbg\" (UID: \"afb60be7-f83c-4dc3-9129-ec66a8b75a55\") " pod="openshift-marketplace/certified-operators-mxkbg" Oct 11 03:55:40 crc kubenswrapper[4967]: I1011 03:55:40.291002 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-db2px\" (UniqueName: \"kubernetes.io/projected/afb60be7-f83c-4dc3-9129-ec66a8b75a55-kube-api-access-db2px\") pod \"certified-operators-mxkbg\" (UID: \"afb60be7-f83c-4dc3-9129-ec66a8b75a55\") " pod="openshift-marketplace/certified-operators-mxkbg" Oct 11 03:55:40 crc kubenswrapper[4967]: I1011 03:55:40.478824 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mxkbg" Oct 11 03:55:40 crc kubenswrapper[4967]: I1011 03:55:40.633796 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-p2xpx"] Oct 11 03:55:40 crc kubenswrapper[4967]: I1011 03:55:40.832717 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="130a69d7-5216-4017-9146-1940a1d58210" path="/var/lib/kubelet/pods/130a69d7-5216-4017-9146-1940a1d58210/volumes" Oct 11 03:55:40 crc kubenswrapper[4967]: I1011 03:55:40.833518 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9509019-37bf-4674-af71-c259393eab0b" path="/var/lib/kubelet/pods/b9509019-37bf-4674-af71-c259393eab0b/volumes" Oct 11 03:55:40 crc kubenswrapper[4967]: I1011 03:55:40.834256 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d05dc46c-e2ec-4c4b-bfed-dd509f93a5df" path="/var/lib/kubelet/pods/d05dc46c-e2ec-4c4b-bfed-dd509f93a5df/volumes" Oct 11 03:55:40 crc kubenswrapper[4967]: I1011 03:55:40.835777 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f70cf6aa-6c00-4507-98c3-4f385cc57522" path="/var/lib/kubelet/pods/f70cf6aa-6c00-4507-98c3-4f385cc57522/volumes" Oct 11 03:55:40 crc kubenswrapper[4967]: I1011 03:55:40.836946 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac" path="/var/lib/kubelet/pods/f92c66af-b0e3-41e4-9ab4-36e5fe4b5fac/volumes" Oct 11 03:55:40 crc kubenswrapper[4967]: I1011 03:55:40.865496 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mxkbg"] Oct 11 03:55:41 crc kubenswrapper[4967]: I1011 03:55:41.437755 4967 generic.go:334] "Generic (PLEG): container finished" podID="065a2523-459f-440e-b5da-5ea5318fd89f" containerID="0032f9362c8fd83b447e4375b5d37e8c42d2bc32650f419eebbf42e599cb3ea5" exitCode=0 Oct 11 03:55:41 crc kubenswrapper[4967]: I1011 03:55:41.437855 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p2xpx" event={"ID":"065a2523-459f-440e-b5da-5ea5318fd89f","Type":"ContainerDied","Data":"0032f9362c8fd83b447e4375b5d37e8c42d2bc32650f419eebbf42e599cb3ea5"} Oct 11 03:55:41 crc kubenswrapper[4967]: I1011 03:55:41.437897 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p2xpx" event={"ID":"065a2523-459f-440e-b5da-5ea5318fd89f","Type":"ContainerStarted","Data":"e41e2ecdcdec67f1b937dfa32cdbb7f0c5fa4dc22283dc87a0fc5a32a7fa312a"} Oct 11 03:55:41 crc kubenswrapper[4967]: I1011 03:55:41.439266 4967 generic.go:334] "Generic (PLEG): container finished" podID="afb60be7-f83c-4dc3-9129-ec66a8b75a55" containerID="411b4766b898e104cf65f8b2bbc4d2c70ac12e1b4d35d3c4469e61fb21c93e66" exitCode=0 Oct 11 03:55:41 crc kubenswrapper[4967]: I1011 03:55:41.439374 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mxkbg" event={"ID":"afb60be7-f83c-4dc3-9129-ec66a8b75a55","Type":"ContainerDied","Data":"411b4766b898e104cf65f8b2bbc4d2c70ac12e1b4d35d3c4469e61fb21c93e66"} Oct 11 03:55:41 crc kubenswrapper[4967]: I1011 03:55:41.439467 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mxkbg" event={"ID":"afb60be7-f83c-4dc3-9129-ec66a8b75a55","Type":"ContainerStarted","Data":"d9a1b9d8130ced2f80565d2191fdca0d8b554b959f76a78bb4d6a1492e6ccbd3"} Oct 11 03:55:42 crc kubenswrapper[4967]: I1011 03:55:42.301711 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dgst6"] Oct 11 03:55:42 crc kubenswrapper[4967]: I1011 03:55:42.303162 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dgst6" Oct 11 03:55:42 crc kubenswrapper[4967]: I1011 03:55:42.305688 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Oct 11 03:55:42 crc kubenswrapper[4967]: I1011 03:55:42.324147 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dgst6"] Oct 11 03:55:42 crc kubenswrapper[4967]: I1011 03:55:42.398353 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f8hhf\" (UniqueName: \"kubernetes.io/projected/cd79944d-93c5-45d4-b68d-5cff92c3d111-kube-api-access-f8hhf\") pod \"community-operators-dgst6\" (UID: \"cd79944d-93c5-45d4-b68d-5cff92c3d111\") " pod="openshift-marketplace/community-operators-dgst6" Oct 11 03:55:42 crc kubenswrapper[4967]: I1011 03:55:42.398400 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd79944d-93c5-45d4-b68d-5cff92c3d111-utilities\") pod \"community-operators-dgst6\" (UID: \"cd79944d-93c5-45d4-b68d-5cff92c3d111\") " pod="openshift-marketplace/community-operators-dgst6" Oct 11 03:55:42 crc kubenswrapper[4967]: I1011 03:55:42.398468 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd79944d-93c5-45d4-b68d-5cff92c3d111-catalog-content\") pod \"community-operators-dgst6\" (UID: \"cd79944d-93c5-45d4-b68d-5cff92c3d111\") " pod="openshift-marketplace/community-operators-dgst6" Oct 11 03:55:42 crc kubenswrapper[4967]: I1011 03:55:42.446214 4967 generic.go:334] "Generic (PLEG): container finished" podID="afb60be7-f83c-4dc3-9129-ec66a8b75a55" containerID="af3827296d0ee25a1d9d7024d918304edef17d6d65f83334665abe2f9ce77d70" exitCode=0 Oct 11 03:55:42 crc kubenswrapper[4967]: I1011 03:55:42.446270 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mxkbg" event={"ID":"afb60be7-f83c-4dc3-9129-ec66a8b75a55","Type":"ContainerDied","Data":"af3827296d0ee25a1d9d7024d918304edef17d6d65f83334665abe2f9ce77d70"} Oct 11 03:55:42 crc kubenswrapper[4967]: I1011 03:55:42.449363 4967 generic.go:334] "Generic (PLEG): container finished" podID="065a2523-459f-440e-b5da-5ea5318fd89f" containerID="fe6ab28af111c6d22a87a162e8cf866cdcd8ad80563792b4b20201ad37206e1d" exitCode=0 Oct 11 03:55:42 crc kubenswrapper[4967]: I1011 03:55:42.449416 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p2xpx" event={"ID":"065a2523-459f-440e-b5da-5ea5318fd89f","Type":"ContainerDied","Data":"fe6ab28af111c6d22a87a162e8cf866cdcd8ad80563792b4b20201ad37206e1d"} Oct 11 03:55:42 crc kubenswrapper[4967]: I1011 03:55:42.499353 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd79944d-93c5-45d4-b68d-5cff92c3d111-catalog-content\") pod \"community-operators-dgst6\" (UID: \"cd79944d-93c5-45d4-b68d-5cff92c3d111\") " pod="openshift-marketplace/community-operators-dgst6" Oct 11 03:55:42 crc kubenswrapper[4967]: I1011 03:55:42.499489 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f8hhf\" (UniqueName: \"kubernetes.io/projected/cd79944d-93c5-45d4-b68d-5cff92c3d111-kube-api-access-f8hhf\") pod \"community-operators-dgst6\" (UID: \"cd79944d-93c5-45d4-b68d-5cff92c3d111\") " pod="openshift-marketplace/community-operators-dgst6" Oct 11 03:55:42 crc kubenswrapper[4967]: I1011 03:55:42.499557 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd79944d-93c5-45d4-b68d-5cff92c3d111-utilities\") pod \"community-operators-dgst6\" (UID: \"cd79944d-93c5-45d4-b68d-5cff92c3d111\") " pod="openshift-marketplace/community-operators-dgst6" Oct 11 03:55:42 crc kubenswrapper[4967]: I1011 03:55:42.499898 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/cd79944d-93c5-45d4-b68d-5cff92c3d111-catalog-content\") pod \"community-operators-dgst6\" (UID: \"cd79944d-93c5-45d4-b68d-5cff92c3d111\") " pod="openshift-marketplace/community-operators-dgst6" Oct 11 03:55:42 crc kubenswrapper[4967]: I1011 03:55:42.500017 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/cd79944d-93c5-45d4-b68d-5cff92c3d111-utilities\") pod \"community-operators-dgst6\" (UID: \"cd79944d-93c5-45d4-b68d-5cff92c3d111\") " pod="openshift-marketplace/community-operators-dgst6" Oct 11 03:55:42 crc kubenswrapper[4967]: I1011 03:55:42.504198 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-khtzc"] Oct 11 03:55:42 crc kubenswrapper[4967]: I1011 03:55:42.505363 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-khtzc" Oct 11 03:55:42 crc kubenswrapper[4967]: I1011 03:55:42.512202 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-khtzc"] Oct 11 03:55:42 crc kubenswrapper[4967]: I1011 03:55:42.512913 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Oct 11 03:55:42 crc kubenswrapper[4967]: I1011 03:55:42.540267 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f8hhf\" (UniqueName: \"kubernetes.io/projected/cd79944d-93c5-45d4-b68d-5cff92c3d111-kube-api-access-f8hhf\") pod \"community-operators-dgst6\" (UID: \"cd79944d-93c5-45d4-b68d-5cff92c3d111\") " pod="openshift-marketplace/community-operators-dgst6" Oct 11 03:55:42 crc kubenswrapper[4967]: I1011 03:55:42.601494 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/553fcde4-86f4-4a03-9650-bb6bd3ee5d0e-utilities\") pod \"redhat-operators-khtzc\" (UID: \"553fcde4-86f4-4a03-9650-bb6bd3ee5d0e\") " pod="openshift-marketplace/redhat-operators-khtzc" Oct 11 03:55:42 crc kubenswrapper[4967]: I1011 03:55:42.601580 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kzz8g\" (UniqueName: \"kubernetes.io/projected/553fcde4-86f4-4a03-9650-bb6bd3ee5d0e-kube-api-access-kzz8g\") pod \"redhat-operators-khtzc\" (UID: \"553fcde4-86f4-4a03-9650-bb6bd3ee5d0e\") " pod="openshift-marketplace/redhat-operators-khtzc" Oct 11 03:55:42 crc kubenswrapper[4967]: I1011 03:55:42.601702 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/553fcde4-86f4-4a03-9650-bb6bd3ee5d0e-catalog-content\") pod \"redhat-operators-khtzc\" (UID: \"553fcde4-86f4-4a03-9650-bb6bd3ee5d0e\") " pod="openshift-marketplace/redhat-operators-khtzc" Oct 11 03:55:42 crc kubenswrapper[4967]: I1011 03:55:42.628370 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dgst6" Oct 11 03:55:42 crc kubenswrapper[4967]: I1011 03:55:42.709195 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/553fcde4-86f4-4a03-9650-bb6bd3ee5d0e-utilities\") pod \"redhat-operators-khtzc\" (UID: \"553fcde4-86f4-4a03-9650-bb6bd3ee5d0e\") " pod="openshift-marketplace/redhat-operators-khtzc" Oct 11 03:55:42 crc kubenswrapper[4967]: I1011 03:55:42.709564 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kzz8g\" (UniqueName: \"kubernetes.io/projected/553fcde4-86f4-4a03-9650-bb6bd3ee5d0e-kube-api-access-kzz8g\") pod \"redhat-operators-khtzc\" (UID: \"553fcde4-86f4-4a03-9650-bb6bd3ee5d0e\") " pod="openshift-marketplace/redhat-operators-khtzc" Oct 11 03:55:42 crc kubenswrapper[4967]: I1011 03:55:42.709750 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/553fcde4-86f4-4a03-9650-bb6bd3ee5d0e-catalog-content\") pod \"redhat-operators-khtzc\" (UID: \"553fcde4-86f4-4a03-9650-bb6bd3ee5d0e\") " pod="openshift-marketplace/redhat-operators-khtzc" Oct 11 03:55:42 crc kubenswrapper[4967]: I1011 03:55:42.710973 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/553fcde4-86f4-4a03-9650-bb6bd3ee5d0e-catalog-content\") pod \"redhat-operators-khtzc\" (UID: \"553fcde4-86f4-4a03-9650-bb6bd3ee5d0e\") " pod="openshift-marketplace/redhat-operators-khtzc" Oct 11 03:55:42 crc kubenswrapper[4967]: I1011 03:55:42.711364 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/553fcde4-86f4-4a03-9650-bb6bd3ee5d0e-utilities\") pod \"redhat-operators-khtzc\" (UID: \"553fcde4-86f4-4a03-9650-bb6bd3ee5d0e\") " pod="openshift-marketplace/redhat-operators-khtzc" Oct 11 03:55:42 crc kubenswrapper[4967]: I1011 03:55:42.732560 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kzz8g\" (UniqueName: \"kubernetes.io/projected/553fcde4-86f4-4a03-9650-bb6bd3ee5d0e-kube-api-access-kzz8g\") pod \"redhat-operators-khtzc\" (UID: \"553fcde4-86f4-4a03-9650-bb6bd3ee5d0e\") " pod="openshift-marketplace/redhat-operators-khtzc" Oct 11 03:55:42 crc kubenswrapper[4967]: I1011 03:55:42.860691 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-khtzc" Oct 11 03:55:43 crc kubenswrapper[4967]: I1011 03:55:43.020696 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dgst6"] Oct 11 03:55:43 crc kubenswrapper[4967]: W1011 03:55:43.034584 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd79944d_93c5_45d4_b68d_5cff92c3d111.slice/crio-beabe611ff1d1d8d2aec8099941262e1938e93cf99ffd7a04bb71fde95d5023d WatchSource:0}: Error finding container beabe611ff1d1d8d2aec8099941262e1938e93cf99ffd7a04bb71fde95d5023d: Status 404 returned error can't find the container with id beabe611ff1d1d8d2aec8099941262e1938e93cf99ffd7a04bb71fde95d5023d Oct 11 03:55:43 crc kubenswrapper[4967]: I1011 03:55:43.264434 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-khtzc"] Oct 11 03:55:43 crc kubenswrapper[4967]: I1011 03:55:43.456257 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-p2xpx" event={"ID":"065a2523-459f-440e-b5da-5ea5318fd89f","Type":"ContainerStarted","Data":"e5802960969084fe3075167164e7bfdf9b282b2065d00ac06d4be6bfa10cae7c"} Oct 11 03:55:43 crc kubenswrapper[4967]: I1011 03:55:43.457623 4967 generic.go:334] "Generic (PLEG): container finished" podID="cd79944d-93c5-45d4-b68d-5cff92c3d111" containerID="43ae41f5b8807949bf8636603ece496a7ec966cebe29dac2dd25c2e20e3d4d8f" exitCode=0 Oct 11 03:55:43 crc kubenswrapper[4967]: I1011 03:55:43.457735 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dgst6" event={"ID":"cd79944d-93c5-45d4-b68d-5cff92c3d111","Type":"ContainerDied","Data":"43ae41f5b8807949bf8636603ece496a7ec966cebe29dac2dd25c2e20e3d4d8f"} Oct 11 03:55:43 crc kubenswrapper[4967]: I1011 03:55:43.457797 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dgst6" event={"ID":"cd79944d-93c5-45d4-b68d-5cff92c3d111","Type":"ContainerStarted","Data":"beabe611ff1d1d8d2aec8099941262e1938e93cf99ffd7a04bb71fde95d5023d"} Oct 11 03:55:43 crc kubenswrapper[4967]: I1011 03:55:43.460332 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mxkbg" event={"ID":"afb60be7-f83c-4dc3-9129-ec66a8b75a55","Type":"ContainerStarted","Data":"d1d8b6c29393fc36e458dd74d0e6e89d02d80d00fb8e921a6bfae0c9696385b2"} Oct 11 03:55:43 crc kubenswrapper[4967]: I1011 03:55:43.461713 4967 generic.go:334] "Generic (PLEG): container finished" podID="553fcde4-86f4-4a03-9650-bb6bd3ee5d0e" containerID="602b81a693ef92f965fe897396b76e5c140ebfecd76410ee78239b0805040c32" exitCode=0 Oct 11 03:55:43 crc kubenswrapper[4967]: I1011 03:55:43.461753 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-khtzc" event={"ID":"553fcde4-86f4-4a03-9650-bb6bd3ee5d0e","Type":"ContainerDied","Data":"602b81a693ef92f965fe897396b76e5c140ebfecd76410ee78239b0805040c32"} Oct 11 03:55:43 crc kubenswrapper[4967]: I1011 03:55:43.461773 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-khtzc" event={"ID":"553fcde4-86f4-4a03-9650-bb6bd3ee5d0e","Type":"ContainerStarted","Data":"0ea2893c3de2cbc3bcd673009f136f959e05f32de044151633751d1bf085ec89"} Oct 11 03:55:43 crc kubenswrapper[4967]: I1011 03:55:43.475883 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-p2xpx" podStartSLOduration=2.994241731 podStartE2EDuration="4.475861953s" podCreationTimestamp="2025-10-11 03:55:39 +0000 UTC" firstStartedPulling="2025-10-11 03:55:41.441642473 +0000 UTC m=+269.404851446" lastFinishedPulling="2025-10-11 03:55:42.923262735 +0000 UTC m=+270.886471668" observedRunningTime="2025-10-11 03:55:43.47260433 +0000 UTC m=+271.435813263" watchObservedRunningTime="2025-10-11 03:55:43.475861953 +0000 UTC m=+271.439070896" Oct 11 03:55:43 crc kubenswrapper[4967]: I1011 03:55:43.488531 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mxkbg" podStartSLOduration=1.8362351609999998 podStartE2EDuration="3.488512583s" podCreationTimestamp="2025-10-11 03:55:40 +0000 UTC" firstStartedPulling="2025-10-11 03:55:41.441189861 +0000 UTC m=+269.404398794" lastFinishedPulling="2025-10-11 03:55:43.093467283 +0000 UTC m=+271.056676216" observedRunningTime="2025-10-11 03:55:43.487476426 +0000 UTC m=+271.450685349" watchObservedRunningTime="2025-10-11 03:55:43.488512583 +0000 UTC m=+271.451721516" Oct 11 03:55:44 crc kubenswrapper[4967]: I1011 03:55:44.468301 4967 generic.go:334] "Generic (PLEG): container finished" podID="cd79944d-93c5-45d4-b68d-5cff92c3d111" containerID="289128e8f1a09fba876f9f69b318d10db2606db6e3adaf48c73e013e44bf46e1" exitCode=0 Oct 11 03:55:44 crc kubenswrapper[4967]: I1011 03:55:44.468368 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dgst6" event={"ID":"cd79944d-93c5-45d4-b68d-5cff92c3d111","Type":"ContainerDied","Data":"289128e8f1a09fba876f9f69b318d10db2606db6e3adaf48c73e013e44bf46e1"} Oct 11 03:55:44 crc kubenswrapper[4967]: I1011 03:55:44.470599 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-khtzc" event={"ID":"553fcde4-86f4-4a03-9650-bb6bd3ee5d0e","Type":"ContainerStarted","Data":"9ffef1fe623f8934f60c15d1378e099791077095116637f72bf553dca51a4813"} Oct 11 03:55:45 crc kubenswrapper[4967]: I1011 03:55:45.478485 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dgst6" event={"ID":"cd79944d-93c5-45d4-b68d-5cff92c3d111","Type":"ContainerStarted","Data":"475b328df535a81ac0931df2db4c208dfcb56671f0904b0f59a465fd6b004e99"} Oct 11 03:55:45 crc kubenswrapper[4967]: I1011 03:55:45.485147 4967 generic.go:334] "Generic (PLEG): container finished" podID="553fcde4-86f4-4a03-9650-bb6bd3ee5d0e" containerID="9ffef1fe623f8934f60c15d1378e099791077095116637f72bf553dca51a4813" exitCode=0 Oct 11 03:55:45 crc kubenswrapper[4967]: I1011 03:55:45.485218 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-khtzc" event={"ID":"553fcde4-86f4-4a03-9650-bb6bd3ee5d0e","Type":"ContainerDied","Data":"9ffef1fe623f8934f60c15d1378e099791077095116637f72bf553dca51a4813"} Oct 11 03:55:45 crc kubenswrapper[4967]: I1011 03:55:45.500171 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dgst6" podStartSLOduration=2.051279437 podStartE2EDuration="3.500148811s" podCreationTimestamp="2025-10-11 03:55:42 +0000 UTC" firstStartedPulling="2025-10-11 03:55:43.45956087 +0000 UTC m=+271.422769793" lastFinishedPulling="2025-10-11 03:55:44.908430234 +0000 UTC m=+272.871639167" observedRunningTime="2025-10-11 03:55:45.497424882 +0000 UTC m=+273.460633825" watchObservedRunningTime="2025-10-11 03:55:45.500148811 +0000 UTC m=+273.463357764" Oct 11 03:55:47 crc kubenswrapper[4967]: I1011 03:55:47.501447 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-khtzc" event={"ID":"553fcde4-86f4-4a03-9650-bb6bd3ee5d0e","Type":"ContainerStarted","Data":"5223b95f63cf6e6202bdf74fe53823f2b4167a1219e6fbc14e38310d199bd3fc"} Oct 11 03:55:47 crc kubenswrapper[4967]: I1011 03:55:47.519963 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-khtzc" podStartSLOduration=2.887243378 podStartE2EDuration="5.519928526s" podCreationTimestamp="2025-10-11 03:55:42 +0000 UTC" firstStartedPulling="2025-10-11 03:55:43.462818372 +0000 UTC m=+271.426027305" lastFinishedPulling="2025-10-11 03:55:46.09550352 +0000 UTC m=+274.058712453" observedRunningTime="2025-10-11 03:55:47.516805577 +0000 UTC m=+275.480014520" watchObservedRunningTime="2025-10-11 03:55:47.519928526 +0000 UTC m=+275.483137489" Oct 11 03:55:50 crc kubenswrapper[4967]: I1011 03:55:50.219916 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-p2xpx" Oct 11 03:55:50 crc kubenswrapper[4967]: I1011 03:55:50.220188 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-p2xpx" Oct 11 03:55:50 crc kubenswrapper[4967]: I1011 03:55:50.302443 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-p2xpx" Oct 11 03:55:50 crc kubenswrapper[4967]: I1011 03:55:50.479354 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-mxkbg" Oct 11 03:55:50 crc kubenswrapper[4967]: I1011 03:55:50.479734 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-mxkbg" Oct 11 03:55:50 crc kubenswrapper[4967]: I1011 03:55:50.543306 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-mxkbg" Oct 11 03:55:50 crc kubenswrapper[4967]: I1011 03:55:50.579117 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-p2xpx" Oct 11 03:55:51 crc kubenswrapper[4967]: I1011 03:55:51.564281 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-mxkbg" Oct 11 03:55:52 crc kubenswrapper[4967]: I1011 03:55:52.629107 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dgst6" Oct 11 03:55:52 crc kubenswrapper[4967]: I1011 03:55:52.630002 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dgst6" Oct 11 03:55:52 crc kubenswrapper[4967]: I1011 03:55:52.676159 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dgst6" Oct 11 03:55:52 crc kubenswrapper[4967]: I1011 03:55:52.861725 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-khtzc" Oct 11 03:55:52 crc kubenswrapper[4967]: I1011 03:55:52.861773 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-khtzc" Oct 11 03:55:52 crc kubenswrapper[4967]: I1011 03:55:52.919542 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-khtzc" Oct 11 03:55:53 crc kubenswrapper[4967]: I1011 03:55:53.598520 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-khtzc" Oct 11 03:55:53 crc kubenswrapper[4967]: I1011 03:55:53.613508 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dgst6" Oct 11 03:56:44 crc kubenswrapper[4967]: I1011 03:56:44.089048 4967 patch_prober.go:28] interesting pod/machine-config-daemon-sgrv2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:56:44 crc kubenswrapper[4967]: I1011 03:56:44.089785 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:57:14 crc kubenswrapper[4967]: I1011 03:57:14.089275 4967 patch_prober.go:28] interesting pod/machine-config-daemon-sgrv2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:57:14 crc kubenswrapper[4967]: I1011 03:57:14.090017 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:57:44 crc kubenswrapper[4967]: I1011 03:57:44.088760 4967 patch_prober.go:28] interesting pod/machine-config-daemon-sgrv2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:57:44 crc kubenswrapper[4967]: I1011 03:57:44.090123 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:57:44 crc kubenswrapper[4967]: I1011 03:57:44.090207 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" Oct 11 03:57:44 crc kubenswrapper[4967]: I1011 03:57:44.091849 4967 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"7cfd4e9315984debe791c2e9396234b602578eb12f2fc553132988f8ec022205"} pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 11 03:57:44 crc kubenswrapper[4967]: I1011 03:57:44.091945 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" containerID="cri-o://7cfd4e9315984debe791c2e9396234b602578eb12f2fc553132988f8ec022205" gracePeriod=600 Oct 11 03:57:44 crc kubenswrapper[4967]: I1011 03:57:44.251592 4967 generic.go:334] "Generic (PLEG): container finished" podID="97c07678-14be-410c-b61f-498cb49bc960" containerID="7cfd4e9315984debe791c2e9396234b602578eb12f2fc553132988f8ec022205" exitCode=0 Oct 11 03:57:44 crc kubenswrapper[4967]: I1011 03:57:44.251633 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" event={"ID":"97c07678-14be-410c-b61f-498cb49bc960","Type":"ContainerDied","Data":"7cfd4e9315984debe791c2e9396234b602578eb12f2fc553132988f8ec022205"} Oct 11 03:57:44 crc kubenswrapper[4967]: I1011 03:57:44.251680 4967 scope.go:117] "RemoveContainer" containerID="a0bb43426a9dd4d6fcebb93069e1ba2b81ea80499280a7c6cb506c7eef1286fb" Oct 11 03:57:45 crc kubenswrapper[4967]: I1011 03:57:45.261774 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" event={"ID":"97c07678-14be-410c-b61f-498cb49bc960","Type":"ContainerStarted","Data":"99f1dbfbe3816d0d177a51d87206e04c54f4faf6dd94db82be3e55ea86cf71f4"} Oct 11 03:59:01 crc kubenswrapper[4967]: I1011 03:59:01.704841 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-9748v"] Oct 11 03:59:01 crc kubenswrapper[4967]: I1011 03:59:01.706048 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-9748v" Oct 11 03:59:01 crc kubenswrapper[4967]: I1011 03:59:01.724436 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-9748v"] Oct 11 03:59:01 crc kubenswrapper[4967]: I1011 03:59:01.733549 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/abc530fb-e286-4240-b7d4-3476bef48b19-trusted-ca\") pod \"image-registry-66df7c8f76-9748v\" (UID: \"abc530fb-e286-4240-b7d4-3476bef48b19\") " pod="openshift-image-registry/image-registry-66df7c8f76-9748v" Oct 11 03:59:01 crc kubenswrapper[4967]: I1011 03:59:01.733641 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q9n4h\" (UniqueName: \"kubernetes.io/projected/abc530fb-e286-4240-b7d4-3476bef48b19-kube-api-access-q9n4h\") pod \"image-registry-66df7c8f76-9748v\" (UID: \"abc530fb-e286-4240-b7d4-3476bef48b19\") " pod="openshift-image-registry/image-registry-66df7c8f76-9748v" Oct 11 03:59:01 crc kubenswrapper[4967]: I1011 03:59:01.733670 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/abc530fb-e286-4240-b7d4-3476bef48b19-registry-certificates\") pod \"image-registry-66df7c8f76-9748v\" (UID: \"abc530fb-e286-4240-b7d4-3476bef48b19\") " pod="openshift-image-registry/image-registry-66df7c8f76-9748v" Oct 11 03:59:01 crc kubenswrapper[4967]: I1011 03:59:01.733692 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/abc530fb-e286-4240-b7d4-3476bef48b19-bound-sa-token\") pod \"image-registry-66df7c8f76-9748v\" (UID: \"abc530fb-e286-4240-b7d4-3476bef48b19\") " pod="openshift-image-registry/image-registry-66df7c8f76-9748v" Oct 11 03:59:01 crc kubenswrapper[4967]: I1011 03:59:01.733718 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/abc530fb-e286-4240-b7d4-3476bef48b19-installation-pull-secrets\") pod \"image-registry-66df7c8f76-9748v\" (UID: \"abc530fb-e286-4240-b7d4-3476bef48b19\") " pod="openshift-image-registry/image-registry-66df7c8f76-9748v" Oct 11 03:59:01 crc kubenswrapper[4967]: I1011 03:59:01.733756 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-9748v\" (UID: \"abc530fb-e286-4240-b7d4-3476bef48b19\") " pod="openshift-image-registry/image-registry-66df7c8f76-9748v" Oct 11 03:59:01 crc kubenswrapper[4967]: I1011 03:59:01.733772 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/abc530fb-e286-4240-b7d4-3476bef48b19-registry-tls\") pod \"image-registry-66df7c8f76-9748v\" (UID: \"abc530fb-e286-4240-b7d4-3476bef48b19\") " pod="openshift-image-registry/image-registry-66df7c8f76-9748v" Oct 11 03:59:01 crc kubenswrapper[4967]: I1011 03:59:01.733797 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/abc530fb-e286-4240-b7d4-3476bef48b19-ca-trust-extracted\") pod \"image-registry-66df7c8f76-9748v\" (UID: \"abc530fb-e286-4240-b7d4-3476bef48b19\") " pod="openshift-image-registry/image-registry-66df7c8f76-9748v" Oct 11 03:59:01 crc kubenswrapper[4967]: I1011 03:59:01.782384 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-9748v\" (UID: \"abc530fb-e286-4240-b7d4-3476bef48b19\") " pod="openshift-image-registry/image-registry-66df7c8f76-9748v" Oct 11 03:59:01 crc kubenswrapper[4967]: I1011 03:59:01.834640 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/abc530fb-e286-4240-b7d4-3476bef48b19-trusted-ca\") pod \"image-registry-66df7c8f76-9748v\" (UID: \"abc530fb-e286-4240-b7d4-3476bef48b19\") " pod="openshift-image-registry/image-registry-66df7c8f76-9748v" Oct 11 03:59:01 crc kubenswrapper[4967]: I1011 03:59:01.834730 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q9n4h\" (UniqueName: \"kubernetes.io/projected/abc530fb-e286-4240-b7d4-3476bef48b19-kube-api-access-q9n4h\") pod \"image-registry-66df7c8f76-9748v\" (UID: \"abc530fb-e286-4240-b7d4-3476bef48b19\") " pod="openshift-image-registry/image-registry-66df7c8f76-9748v" Oct 11 03:59:01 crc kubenswrapper[4967]: I1011 03:59:01.834762 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/abc530fb-e286-4240-b7d4-3476bef48b19-registry-certificates\") pod \"image-registry-66df7c8f76-9748v\" (UID: \"abc530fb-e286-4240-b7d4-3476bef48b19\") " pod="openshift-image-registry/image-registry-66df7c8f76-9748v" Oct 11 03:59:01 crc kubenswrapper[4967]: I1011 03:59:01.834783 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/abc530fb-e286-4240-b7d4-3476bef48b19-bound-sa-token\") pod \"image-registry-66df7c8f76-9748v\" (UID: \"abc530fb-e286-4240-b7d4-3476bef48b19\") " pod="openshift-image-registry/image-registry-66df7c8f76-9748v" Oct 11 03:59:01 crc kubenswrapper[4967]: I1011 03:59:01.834808 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/abc530fb-e286-4240-b7d4-3476bef48b19-installation-pull-secrets\") pod \"image-registry-66df7c8f76-9748v\" (UID: \"abc530fb-e286-4240-b7d4-3476bef48b19\") " pod="openshift-image-registry/image-registry-66df7c8f76-9748v" Oct 11 03:59:01 crc kubenswrapper[4967]: I1011 03:59:01.834842 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/abc530fb-e286-4240-b7d4-3476bef48b19-registry-tls\") pod \"image-registry-66df7c8f76-9748v\" (UID: \"abc530fb-e286-4240-b7d4-3476bef48b19\") " pod="openshift-image-registry/image-registry-66df7c8f76-9748v" Oct 11 03:59:01 crc kubenswrapper[4967]: I1011 03:59:01.834875 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/abc530fb-e286-4240-b7d4-3476bef48b19-ca-trust-extracted\") pod \"image-registry-66df7c8f76-9748v\" (UID: \"abc530fb-e286-4240-b7d4-3476bef48b19\") " pod="openshift-image-registry/image-registry-66df7c8f76-9748v" Oct 11 03:59:01 crc kubenswrapper[4967]: I1011 03:59:01.835451 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/abc530fb-e286-4240-b7d4-3476bef48b19-ca-trust-extracted\") pod \"image-registry-66df7c8f76-9748v\" (UID: \"abc530fb-e286-4240-b7d4-3476bef48b19\") " pod="openshift-image-registry/image-registry-66df7c8f76-9748v" Oct 11 03:59:01 crc kubenswrapper[4967]: I1011 03:59:01.836003 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/abc530fb-e286-4240-b7d4-3476bef48b19-trusted-ca\") pod \"image-registry-66df7c8f76-9748v\" (UID: \"abc530fb-e286-4240-b7d4-3476bef48b19\") " pod="openshift-image-registry/image-registry-66df7c8f76-9748v" Oct 11 03:59:01 crc kubenswrapper[4967]: I1011 03:59:01.837176 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/abc530fb-e286-4240-b7d4-3476bef48b19-registry-certificates\") pod \"image-registry-66df7c8f76-9748v\" (UID: \"abc530fb-e286-4240-b7d4-3476bef48b19\") " pod="openshift-image-registry/image-registry-66df7c8f76-9748v" Oct 11 03:59:01 crc kubenswrapper[4967]: I1011 03:59:01.841841 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/abc530fb-e286-4240-b7d4-3476bef48b19-registry-tls\") pod \"image-registry-66df7c8f76-9748v\" (UID: \"abc530fb-e286-4240-b7d4-3476bef48b19\") " pod="openshift-image-registry/image-registry-66df7c8f76-9748v" Oct 11 03:59:01 crc kubenswrapper[4967]: I1011 03:59:01.844699 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/abc530fb-e286-4240-b7d4-3476bef48b19-installation-pull-secrets\") pod \"image-registry-66df7c8f76-9748v\" (UID: \"abc530fb-e286-4240-b7d4-3476bef48b19\") " pod="openshift-image-registry/image-registry-66df7c8f76-9748v" Oct 11 03:59:01 crc kubenswrapper[4967]: I1011 03:59:01.851936 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q9n4h\" (UniqueName: \"kubernetes.io/projected/abc530fb-e286-4240-b7d4-3476bef48b19-kube-api-access-q9n4h\") pod \"image-registry-66df7c8f76-9748v\" (UID: \"abc530fb-e286-4240-b7d4-3476bef48b19\") " pod="openshift-image-registry/image-registry-66df7c8f76-9748v" Oct 11 03:59:01 crc kubenswrapper[4967]: I1011 03:59:01.852717 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/abc530fb-e286-4240-b7d4-3476bef48b19-bound-sa-token\") pod \"image-registry-66df7c8f76-9748v\" (UID: \"abc530fb-e286-4240-b7d4-3476bef48b19\") " pod="openshift-image-registry/image-registry-66df7c8f76-9748v" Oct 11 03:59:02 crc kubenswrapper[4967]: I1011 03:59:02.021878 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-9748v" Oct 11 03:59:02 crc kubenswrapper[4967]: I1011 03:59:02.222622 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-9748v"] Oct 11 03:59:02 crc kubenswrapper[4967]: I1011 03:59:02.754230 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-9748v" event={"ID":"abc530fb-e286-4240-b7d4-3476bef48b19","Type":"ContainerStarted","Data":"afc203f990764f3717b859ad0e507c52429425ac9e13696f0d2a4c7b78313a2a"} Oct 11 03:59:02 crc kubenswrapper[4967]: I1011 03:59:02.754292 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-9748v" event={"ID":"abc530fb-e286-4240-b7d4-3476bef48b19","Type":"ContainerStarted","Data":"889e568f4ed1e657d346275a52d4cccc89a549d8e0b684b1ec72914b92f25c73"} Oct 11 03:59:02 crc kubenswrapper[4967]: I1011 03:59:02.754339 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-9748v" Oct 11 03:59:02 crc kubenswrapper[4967]: I1011 03:59:02.779130 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-9748v" podStartSLOduration=1.779060477 podStartE2EDuration="1.779060477s" podCreationTimestamp="2025-10-11 03:59:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 03:59:02.775661789 +0000 UTC m=+470.738870752" watchObservedRunningTime="2025-10-11 03:59:02.779060477 +0000 UTC m=+470.742269410" Oct 11 03:59:22 crc kubenswrapper[4967]: I1011 03:59:22.033463 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-9748v" Oct 11 03:59:22 crc kubenswrapper[4967]: I1011 03:59:22.106880 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hk2cx"] Oct 11 03:59:44 crc kubenswrapper[4967]: I1011 03:59:44.089715 4967 patch_prober.go:28] interesting pod/machine-config-daemon-sgrv2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 03:59:44 crc kubenswrapper[4967]: I1011 03:59:44.090480 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 03:59:47 crc kubenswrapper[4967]: I1011 03:59:47.165860 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" podUID="6b6ccb8c-d307-430b-9232-70497214c0bf" containerName="registry" containerID="cri-o://dfdb4b8a671e8fd016f27b7f407b2bbae9a0e9ba57d72096c8669b6263dffa3a" gracePeriod=30 Oct 11 03:59:47 crc kubenswrapper[4967]: I1011 03:59:47.601619 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:59:47 crc kubenswrapper[4967]: I1011 03:59:47.716271 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6b6ccb8c-d307-430b-9232-70497214c0bf-registry-tls\") pod \"6b6ccb8c-d307-430b-9232-70497214c0bf\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " Oct 11 03:59:47 crc kubenswrapper[4967]: I1011 03:59:47.716329 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6b6ccb8c-d307-430b-9232-70497214c0bf-ca-trust-extracted\") pod \"6b6ccb8c-d307-430b-9232-70497214c0bf\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " Oct 11 03:59:47 crc kubenswrapper[4967]: I1011 03:59:47.716356 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6b6ccb8c-d307-430b-9232-70497214c0bf-bound-sa-token\") pod \"6b6ccb8c-d307-430b-9232-70497214c0bf\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " Oct 11 03:59:47 crc kubenswrapper[4967]: I1011 03:59:47.716385 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-24f2z\" (UniqueName: \"kubernetes.io/projected/6b6ccb8c-d307-430b-9232-70497214c0bf-kube-api-access-24f2z\") pod \"6b6ccb8c-d307-430b-9232-70497214c0bf\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " Oct 11 03:59:47 crc kubenswrapper[4967]: I1011 03:59:47.716497 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6b6ccb8c-d307-430b-9232-70497214c0bf-installation-pull-secrets\") pod \"6b6ccb8c-d307-430b-9232-70497214c0bf\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " Oct 11 03:59:47 crc kubenswrapper[4967]: I1011 03:59:47.716521 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6b6ccb8c-d307-430b-9232-70497214c0bf-registry-certificates\") pod \"6b6ccb8c-d307-430b-9232-70497214c0bf\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " Oct 11 03:59:47 crc kubenswrapper[4967]: I1011 03:59:47.716673 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"6b6ccb8c-d307-430b-9232-70497214c0bf\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " Oct 11 03:59:47 crc kubenswrapper[4967]: I1011 03:59:47.716716 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6b6ccb8c-d307-430b-9232-70497214c0bf-trusted-ca\") pod \"6b6ccb8c-d307-430b-9232-70497214c0bf\" (UID: \"6b6ccb8c-d307-430b-9232-70497214c0bf\") " Oct 11 03:59:47 crc kubenswrapper[4967]: I1011 03:59:47.717612 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b6ccb8c-d307-430b-9232-70497214c0bf-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "6b6ccb8c-d307-430b-9232-70497214c0bf" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:59:47 crc kubenswrapper[4967]: I1011 03:59:47.717996 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6b6ccb8c-d307-430b-9232-70497214c0bf-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "6b6ccb8c-d307-430b-9232-70497214c0bf" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 03:59:47 crc kubenswrapper[4967]: I1011 03:59:47.723260 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6b6ccb8c-d307-430b-9232-70497214c0bf-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "6b6ccb8c-d307-430b-9232-70497214c0bf" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 03:59:47 crc kubenswrapper[4967]: I1011 03:59:47.723684 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b6ccb8c-d307-430b-9232-70497214c0bf-kube-api-access-24f2z" (OuterVolumeSpecName: "kube-api-access-24f2z") pod "6b6ccb8c-d307-430b-9232-70497214c0bf" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf"). InnerVolumeSpecName "kube-api-access-24f2z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:59:47 crc kubenswrapper[4967]: I1011 03:59:47.724569 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b6ccb8c-d307-430b-9232-70497214c0bf-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "6b6ccb8c-d307-430b-9232-70497214c0bf" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:59:47 crc kubenswrapper[4967]: I1011 03:59:47.730732 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6b6ccb8c-d307-430b-9232-70497214c0bf-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "6b6ccb8c-d307-430b-9232-70497214c0bf" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 03:59:47 crc kubenswrapper[4967]: I1011 03:59:47.731594 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "6b6ccb8c-d307-430b-9232-70497214c0bf" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Oct 11 03:59:47 crc kubenswrapper[4967]: I1011 03:59:47.732224 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6b6ccb8c-d307-430b-9232-70497214c0bf-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "6b6ccb8c-d307-430b-9232-70497214c0bf" (UID: "6b6ccb8c-d307-430b-9232-70497214c0bf"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 03:59:47 crc kubenswrapper[4967]: I1011 03:59:47.817808 4967 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/6b6ccb8c-d307-430b-9232-70497214c0bf-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Oct 11 03:59:47 crc kubenswrapper[4967]: I1011 03:59:47.817863 4967 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/6b6ccb8c-d307-430b-9232-70497214c0bf-registry-certificates\") on node \"crc\" DevicePath \"\"" Oct 11 03:59:47 crc kubenswrapper[4967]: I1011 03:59:47.817883 4967 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/6b6ccb8c-d307-430b-9232-70497214c0bf-trusted-ca\") on node \"crc\" DevicePath \"\"" Oct 11 03:59:47 crc kubenswrapper[4967]: I1011 03:59:47.817900 4967 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/6b6ccb8c-d307-430b-9232-70497214c0bf-registry-tls\") on node \"crc\" DevicePath \"\"" Oct 11 03:59:47 crc kubenswrapper[4967]: I1011 03:59:47.817918 4967 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/6b6ccb8c-d307-430b-9232-70497214c0bf-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Oct 11 03:59:47 crc kubenswrapper[4967]: I1011 03:59:47.817935 4967 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/6b6ccb8c-d307-430b-9232-70497214c0bf-bound-sa-token\") on node \"crc\" DevicePath \"\"" Oct 11 03:59:47 crc kubenswrapper[4967]: I1011 03:59:47.817954 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-24f2z\" (UniqueName: \"kubernetes.io/projected/6b6ccb8c-d307-430b-9232-70497214c0bf-kube-api-access-24f2z\") on node \"crc\" DevicePath \"\"" Oct 11 03:59:48 crc kubenswrapper[4967]: I1011 03:59:48.033683 4967 generic.go:334] "Generic (PLEG): container finished" podID="6b6ccb8c-d307-430b-9232-70497214c0bf" containerID="dfdb4b8a671e8fd016f27b7f407b2bbae9a0e9ba57d72096c8669b6263dffa3a" exitCode=0 Oct 11 03:59:48 crc kubenswrapper[4967]: I1011 03:59:48.033751 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" Oct 11 03:59:48 crc kubenswrapper[4967]: I1011 03:59:48.033764 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" event={"ID":"6b6ccb8c-d307-430b-9232-70497214c0bf","Type":"ContainerDied","Data":"dfdb4b8a671e8fd016f27b7f407b2bbae9a0e9ba57d72096c8669b6263dffa3a"} Oct 11 03:59:48 crc kubenswrapper[4967]: I1011 03:59:48.033795 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-hk2cx" event={"ID":"6b6ccb8c-d307-430b-9232-70497214c0bf","Type":"ContainerDied","Data":"162a69a1228fdca679e50de3ce6c6a391f937a475b638b6c90bc6d2d053c5a20"} Oct 11 03:59:48 crc kubenswrapper[4967]: I1011 03:59:48.033814 4967 scope.go:117] "RemoveContainer" containerID="dfdb4b8a671e8fd016f27b7f407b2bbae9a0e9ba57d72096c8669b6263dffa3a" Oct 11 03:59:48 crc kubenswrapper[4967]: I1011 03:59:48.051910 4967 scope.go:117] "RemoveContainer" containerID="dfdb4b8a671e8fd016f27b7f407b2bbae9a0e9ba57d72096c8669b6263dffa3a" Oct 11 03:59:48 crc kubenswrapper[4967]: E1011 03:59:48.052951 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"dfdb4b8a671e8fd016f27b7f407b2bbae9a0e9ba57d72096c8669b6263dffa3a\": container with ID starting with dfdb4b8a671e8fd016f27b7f407b2bbae9a0e9ba57d72096c8669b6263dffa3a not found: ID does not exist" containerID="dfdb4b8a671e8fd016f27b7f407b2bbae9a0e9ba57d72096c8669b6263dffa3a" Oct 11 03:59:48 crc kubenswrapper[4967]: I1011 03:59:48.053052 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"dfdb4b8a671e8fd016f27b7f407b2bbae9a0e9ba57d72096c8669b6263dffa3a"} err="failed to get container status \"dfdb4b8a671e8fd016f27b7f407b2bbae9a0e9ba57d72096c8669b6263dffa3a\": rpc error: code = NotFound desc = could not find container \"dfdb4b8a671e8fd016f27b7f407b2bbae9a0e9ba57d72096c8669b6263dffa3a\": container with ID starting with dfdb4b8a671e8fd016f27b7f407b2bbae9a0e9ba57d72096c8669b6263dffa3a not found: ID does not exist" Oct 11 03:59:48 crc kubenswrapper[4967]: I1011 03:59:48.079054 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hk2cx"] Oct 11 03:59:48 crc kubenswrapper[4967]: I1011 03:59:48.081838 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-hk2cx"] Oct 11 03:59:48 crc kubenswrapper[4967]: I1011 03:59:48.824680 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6b6ccb8c-d307-430b-9232-70497214c0bf" path="/var/lib/kubelet/pods/6b6ccb8c-d307-430b-9232-70497214c0bf/volumes" Oct 11 04:00:00 crc kubenswrapper[4967]: I1011 04:00:00.158411 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335920-vj67q"] Oct 11 04:00:00 crc kubenswrapper[4967]: E1011 04:00:00.159465 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6b6ccb8c-d307-430b-9232-70497214c0bf" containerName="registry" Oct 11 04:00:00 crc kubenswrapper[4967]: I1011 04:00:00.159487 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="6b6ccb8c-d307-430b-9232-70497214c0bf" containerName="registry" Oct 11 04:00:00 crc kubenswrapper[4967]: I1011 04:00:00.159680 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="6b6ccb8c-d307-430b-9232-70497214c0bf" containerName="registry" Oct 11 04:00:00 crc kubenswrapper[4967]: I1011 04:00:00.164517 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335920-vj67q" Oct 11 04:00:00 crc kubenswrapper[4967]: I1011 04:00:00.164787 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335920-vj67q"] Oct 11 04:00:00 crc kubenswrapper[4967]: I1011 04:00:00.167734 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 11 04:00:00 crc kubenswrapper[4967]: I1011 04:00:00.168579 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 11 04:00:00 crc kubenswrapper[4967]: I1011 04:00:00.284647 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/83eafe3f-024d-49ba-aead-844a86ae7f0e-secret-volume\") pod \"collect-profiles-29335920-vj67q\" (UID: \"83eafe3f-024d-49ba-aead-844a86ae7f0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335920-vj67q" Oct 11 04:00:00 crc kubenswrapper[4967]: I1011 04:00:00.285012 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/83eafe3f-024d-49ba-aead-844a86ae7f0e-config-volume\") pod \"collect-profiles-29335920-vj67q\" (UID: \"83eafe3f-024d-49ba-aead-844a86ae7f0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335920-vj67q" Oct 11 04:00:00 crc kubenswrapper[4967]: I1011 04:00:00.285056 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pwjl8\" (UniqueName: \"kubernetes.io/projected/83eafe3f-024d-49ba-aead-844a86ae7f0e-kube-api-access-pwjl8\") pod \"collect-profiles-29335920-vj67q\" (UID: \"83eafe3f-024d-49ba-aead-844a86ae7f0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335920-vj67q" Oct 11 04:00:00 crc kubenswrapper[4967]: I1011 04:00:00.386652 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/83eafe3f-024d-49ba-aead-844a86ae7f0e-config-volume\") pod \"collect-profiles-29335920-vj67q\" (UID: \"83eafe3f-024d-49ba-aead-844a86ae7f0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335920-vj67q" Oct 11 04:00:00 crc kubenswrapper[4967]: I1011 04:00:00.386706 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pwjl8\" (UniqueName: \"kubernetes.io/projected/83eafe3f-024d-49ba-aead-844a86ae7f0e-kube-api-access-pwjl8\") pod \"collect-profiles-29335920-vj67q\" (UID: \"83eafe3f-024d-49ba-aead-844a86ae7f0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335920-vj67q" Oct 11 04:00:00 crc kubenswrapper[4967]: I1011 04:00:00.386745 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/83eafe3f-024d-49ba-aead-844a86ae7f0e-secret-volume\") pod \"collect-profiles-29335920-vj67q\" (UID: \"83eafe3f-024d-49ba-aead-844a86ae7f0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335920-vj67q" Oct 11 04:00:00 crc kubenswrapper[4967]: I1011 04:00:00.387930 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/83eafe3f-024d-49ba-aead-844a86ae7f0e-config-volume\") pod \"collect-profiles-29335920-vj67q\" (UID: \"83eafe3f-024d-49ba-aead-844a86ae7f0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335920-vj67q" Oct 11 04:00:00 crc kubenswrapper[4967]: I1011 04:00:00.393755 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/83eafe3f-024d-49ba-aead-844a86ae7f0e-secret-volume\") pod \"collect-profiles-29335920-vj67q\" (UID: \"83eafe3f-024d-49ba-aead-844a86ae7f0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335920-vj67q" Oct 11 04:00:00 crc kubenswrapper[4967]: I1011 04:00:00.412330 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pwjl8\" (UniqueName: \"kubernetes.io/projected/83eafe3f-024d-49ba-aead-844a86ae7f0e-kube-api-access-pwjl8\") pod \"collect-profiles-29335920-vj67q\" (UID: \"83eafe3f-024d-49ba-aead-844a86ae7f0e\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335920-vj67q" Oct 11 04:00:00 crc kubenswrapper[4967]: I1011 04:00:00.493644 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335920-vj67q" Oct 11 04:00:00 crc kubenswrapper[4967]: I1011 04:00:00.678095 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335920-vj67q"] Oct 11 04:00:01 crc kubenswrapper[4967]: I1011 04:00:01.115666 4967 generic.go:334] "Generic (PLEG): container finished" podID="83eafe3f-024d-49ba-aead-844a86ae7f0e" containerID="5b8f2fb648002efcbc9da9597bf59978e6fe42d25709c50de1de08f9a003306b" exitCode=0 Oct 11 04:00:01 crc kubenswrapper[4967]: I1011 04:00:01.115736 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335920-vj67q" event={"ID":"83eafe3f-024d-49ba-aead-844a86ae7f0e","Type":"ContainerDied","Data":"5b8f2fb648002efcbc9da9597bf59978e6fe42d25709c50de1de08f9a003306b"} Oct 11 04:00:01 crc kubenswrapper[4967]: I1011 04:00:01.115807 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335920-vj67q" event={"ID":"83eafe3f-024d-49ba-aead-844a86ae7f0e","Type":"ContainerStarted","Data":"aac70a679da6956e94c05b25dca6a4d7dab2abb352b510bdae307138b3969690"} Oct 11 04:00:02 crc kubenswrapper[4967]: I1011 04:00:02.375574 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335920-vj67q" Oct 11 04:00:02 crc kubenswrapper[4967]: I1011 04:00:02.510497 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pwjl8\" (UniqueName: \"kubernetes.io/projected/83eafe3f-024d-49ba-aead-844a86ae7f0e-kube-api-access-pwjl8\") pod \"83eafe3f-024d-49ba-aead-844a86ae7f0e\" (UID: \"83eafe3f-024d-49ba-aead-844a86ae7f0e\") " Oct 11 04:00:02 crc kubenswrapper[4967]: I1011 04:00:02.510612 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/83eafe3f-024d-49ba-aead-844a86ae7f0e-secret-volume\") pod \"83eafe3f-024d-49ba-aead-844a86ae7f0e\" (UID: \"83eafe3f-024d-49ba-aead-844a86ae7f0e\") " Oct 11 04:00:02 crc kubenswrapper[4967]: I1011 04:00:02.510750 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/83eafe3f-024d-49ba-aead-844a86ae7f0e-config-volume\") pod \"83eafe3f-024d-49ba-aead-844a86ae7f0e\" (UID: \"83eafe3f-024d-49ba-aead-844a86ae7f0e\") " Oct 11 04:00:02 crc kubenswrapper[4967]: I1011 04:00:02.511950 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/83eafe3f-024d-49ba-aead-844a86ae7f0e-config-volume" (OuterVolumeSpecName: "config-volume") pod "83eafe3f-024d-49ba-aead-844a86ae7f0e" (UID: "83eafe3f-024d-49ba-aead-844a86ae7f0e"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:00:02 crc kubenswrapper[4967]: I1011 04:00:02.516021 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83eafe3f-024d-49ba-aead-844a86ae7f0e-kube-api-access-pwjl8" (OuterVolumeSpecName: "kube-api-access-pwjl8") pod "83eafe3f-024d-49ba-aead-844a86ae7f0e" (UID: "83eafe3f-024d-49ba-aead-844a86ae7f0e"). InnerVolumeSpecName "kube-api-access-pwjl8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:00:02 crc kubenswrapper[4967]: I1011 04:00:02.516564 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83eafe3f-024d-49ba-aead-844a86ae7f0e-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "83eafe3f-024d-49ba-aead-844a86ae7f0e" (UID: "83eafe3f-024d-49ba-aead-844a86ae7f0e"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:00:02 crc kubenswrapper[4967]: I1011 04:00:02.612758 4967 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/83eafe3f-024d-49ba-aead-844a86ae7f0e-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 11 04:00:02 crc kubenswrapper[4967]: I1011 04:00:02.612828 4967 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/83eafe3f-024d-49ba-aead-844a86ae7f0e-config-volume\") on node \"crc\" DevicePath \"\"" Oct 11 04:00:02 crc kubenswrapper[4967]: I1011 04:00:02.612852 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pwjl8\" (UniqueName: \"kubernetes.io/projected/83eafe3f-024d-49ba-aead-844a86ae7f0e-kube-api-access-pwjl8\") on node \"crc\" DevicePath \"\"" Oct 11 04:00:03 crc kubenswrapper[4967]: I1011 04:00:03.128525 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335920-vj67q" event={"ID":"83eafe3f-024d-49ba-aead-844a86ae7f0e","Type":"ContainerDied","Data":"aac70a679da6956e94c05b25dca6a4d7dab2abb352b510bdae307138b3969690"} Oct 11 04:00:03 crc kubenswrapper[4967]: I1011 04:00:03.128560 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="aac70a679da6956e94c05b25dca6a4d7dab2abb352b510bdae307138b3969690" Oct 11 04:00:03 crc kubenswrapper[4967]: I1011 04:00:03.128779 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335920-vj67q" Oct 11 04:00:14 crc kubenswrapper[4967]: I1011 04:00:14.088757 4967 patch_prober.go:28] interesting pod/machine-config-daemon-sgrv2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:00:14 crc kubenswrapper[4967]: I1011 04:00:14.089549 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:00:44 crc kubenswrapper[4967]: I1011 04:00:44.089603 4967 patch_prober.go:28] interesting pod/machine-config-daemon-sgrv2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:00:44 crc kubenswrapper[4967]: I1011 04:00:44.090479 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:00:44 crc kubenswrapper[4967]: I1011 04:00:44.090569 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" Oct 11 04:00:44 crc kubenswrapper[4967]: I1011 04:00:44.091614 4967 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"99f1dbfbe3816d0d177a51d87206e04c54f4faf6dd94db82be3e55ea86cf71f4"} pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 11 04:00:44 crc kubenswrapper[4967]: I1011 04:00:44.091857 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" containerID="cri-o://99f1dbfbe3816d0d177a51d87206e04c54f4faf6dd94db82be3e55ea86cf71f4" gracePeriod=600 Oct 11 04:00:44 crc kubenswrapper[4967]: I1011 04:00:44.408136 4967 generic.go:334] "Generic (PLEG): container finished" podID="97c07678-14be-410c-b61f-498cb49bc960" containerID="99f1dbfbe3816d0d177a51d87206e04c54f4faf6dd94db82be3e55ea86cf71f4" exitCode=0 Oct 11 04:00:44 crc kubenswrapper[4967]: I1011 04:00:44.408255 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" event={"ID":"97c07678-14be-410c-b61f-498cb49bc960","Type":"ContainerDied","Data":"99f1dbfbe3816d0d177a51d87206e04c54f4faf6dd94db82be3e55ea86cf71f4"} Oct 11 04:00:44 crc kubenswrapper[4967]: I1011 04:00:44.408614 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" event={"ID":"97c07678-14be-410c-b61f-498cb49bc960","Type":"ContainerStarted","Data":"fab92459036250055aee13eec5e67ffc42b24499db86755378bf85f3ee4b07d2"} Oct 11 04:00:44 crc kubenswrapper[4967]: I1011 04:00:44.408649 4967 scope.go:117] "RemoveContainer" containerID="7cfd4e9315984debe791c2e9396234b602578eb12f2fc553132988f8ec022205" Oct 11 04:01:09 crc kubenswrapper[4967]: I1011 04:01:09.425180 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-ncqxv"] Oct 11 04:01:09 crc kubenswrapper[4967]: E1011 04:01:09.425945 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83eafe3f-024d-49ba-aead-844a86ae7f0e" containerName="collect-profiles" Oct 11 04:01:09 crc kubenswrapper[4967]: I1011 04:01:09.425958 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="83eafe3f-024d-49ba-aead-844a86ae7f0e" containerName="collect-profiles" Oct 11 04:01:09 crc kubenswrapper[4967]: I1011 04:01:09.426092 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="83eafe3f-024d-49ba-aead-844a86ae7f0e" containerName="collect-profiles" Oct 11 04:01:09 crc kubenswrapper[4967]: I1011 04:01:09.426505 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-ncqxv" Oct 11 04:01:09 crc kubenswrapper[4967]: I1011 04:01:09.428516 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Oct 11 04:01:09 crc kubenswrapper[4967]: I1011 04:01:09.428945 4967 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-swvld" Oct 11 04:01:09 crc kubenswrapper[4967]: I1011 04:01:09.429106 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Oct 11 04:01:09 crc kubenswrapper[4967]: I1011 04:01:09.437774 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-ncqxv"] Oct 11 04:01:09 crc kubenswrapper[4967]: I1011 04:01:09.447884 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-5b446d88c5-ml5f9"] Oct 11 04:01:09 crc kubenswrapper[4967]: I1011 04:01:09.448735 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-ml5f9" Oct 11 04:01:09 crc kubenswrapper[4967]: I1011 04:01:09.451705 4967 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-8vx4j" Oct 11 04:01:09 crc kubenswrapper[4967]: I1011 04:01:09.453119 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-ml5f9"] Oct 11 04:01:09 crc kubenswrapper[4967]: I1011 04:01:09.461628 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-xmk8c"] Oct 11 04:01:09 crc kubenswrapper[4967]: I1011 04:01:09.462918 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-xmk8c" Oct 11 04:01:09 crc kubenswrapper[4967]: I1011 04:01:09.467648 4967 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-g4ncc" Oct 11 04:01:09 crc kubenswrapper[4967]: I1011 04:01:09.474206 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-xmk8c"] Oct 11 04:01:09 crc kubenswrapper[4967]: I1011 04:01:09.530400 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vh4bt\" (UniqueName: \"kubernetes.io/projected/822cb861-f537-4669-8fc9-168f052a4b5d-kube-api-access-vh4bt\") pod \"cert-manager-cainjector-7f985d654d-ncqxv\" (UID: \"822cb861-f537-4669-8fc9-168f052a4b5d\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-ncqxv" Oct 11 04:01:09 crc kubenswrapper[4967]: I1011 04:01:09.632134 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xc28v\" (UniqueName: \"kubernetes.io/projected/be6fdb4b-abb9-4e76-be7e-0c7c579d3bf8-kube-api-access-xc28v\") pod \"cert-manager-5b446d88c5-ml5f9\" (UID: \"be6fdb4b-abb9-4e76-be7e-0c7c579d3bf8\") " pod="cert-manager/cert-manager-5b446d88c5-ml5f9" Oct 11 04:01:09 crc kubenswrapper[4967]: I1011 04:01:09.632249 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vh4bt\" (UniqueName: \"kubernetes.io/projected/822cb861-f537-4669-8fc9-168f052a4b5d-kube-api-access-vh4bt\") pod \"cert-manager-cainjector-7f985d654d-ncqxv\" (UID: \"822cb861-f537-4669-8fc9-168f052a4b5d\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-ncqxv" Oct 11 04:01:09 crc kubenswrapper[4967]: I1011 04:01:09.632312 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rksr\" (UniqueName: \"kubernetes.io/projected/ac43a6d2-5018-4d03-9a26-5da4f2619403-kube-api-access-4rksr\") pod \"cert-manager-webhook-5655c58dd6-xmk8c\" (UID: \"ac43a6d2-5018-4d03-9a26-5da4f2619403\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-xmk8c" Oct 11 04:01:09 crc kubenswrapper[4967]: I1011 04:01:09.652034 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vh4bt\" (UniqueName: \"kubernetes.io/projected/822cb861-f537-4669-8fc9-168f052a4b5d-kube-api-access-vh4bt\") pod \"cert-manager-cainjector-7f985d654d-ncqxv\" (UID: \"822cb861-f537-4669-8fc9-168f052a4b5d\") " pod="cert-manager/cert-manager-cainjector-7f985d654d-ncqxv" Oct 11 04:01:09 crc kubenswrapper[4967]: I1011 04:01:09.733252 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rksr\" (UniqueName: \"kubernetes.io/projected/ac43a6d2-5018-4d03-9a26-5da4f2619403-kube-api-access-4rksr\") pod \"cert-manager-webhook-5655c58dd6-xmk8c\" (UID: \"ac43a6d2-5018-4d03-9a26-5da4f2619403\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-xmk8c" Oct 11 04:01:09 crc kubenswrapper[4967]: I1011 04:01:09.733326 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xc28v\" (UniqueName: \"kubernetes.io/projected/be6fdb4b-abb9-4e76-be7e-0c7c579d3bf8-kube-api-access-xc28v\") pod \"cert-manager-5b446d88c5-ml5f9\" (UID: \"be6fdb4b-abb9-4e76-be7e-0c7c579d3bf8\") " pod="cert-manager/cert-manager-5b446d88c5-ml5f9" Oct 11 04:01:09 crc kubenswrapper[4967]: I1011 04:01:09.742411 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-7f985d654d-ncqxv" Oct 11 04:01:09 crc kubenswrapper[4967]: I1011 04:01:09.751352 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rksr\" (UniqueName: \"kubernetes.io/projected/ac43a6d2-5018-4d03-9a26-5da4f2619403-kube-api-access-4rksr\") pod \"cert-manager-webhook-5655c58dd6-xmk8c\" (UID: \"ac43a6d2-5018-4d03-9a26-5da4f2619403\") " pod="cert-manager/cert-manager-webhook-5655c58dd6-xmk8c" Oct 11 04:01:09 crc kubenswrapper[4967]: I1011 04:01:09.751526 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xc28v\" (UniqueName: \"kubernetes.io/projected/be6fdb4b-abb9-4e76-be7e-0c7c579d3bf8-kube-api-access-xc28v\") pod \"cert-manager-5b446d88c5-ml5f9\" (UID: \"be6fdb4b-abb9-4e76-be7e-0c7c579d3bf8\") " pod="cert-manager/cert-manager-5b446d88c5-ml5f9" Oct 11 04:01:09 crc kubenswrapper[4967]: I1011 04:01:09.772595 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-5b446d88c5-ml5f9" Oct 11 04:01:09 crc kubenswrapper[4967]: I1011 04:01:09.781634 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-5655c58dd6-xmk8c" Oct 11 04:01:09 crc kubenswrapper[4967]: I1011 04:01:09.987280 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-5b446d88c5-ml5f9"] Oct 11 04:01:10 crc kubenswrapper[4967]: I1011 04:01:09.999860 4967 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 11 04:01:10 crc kubenswrapper[4967]: I1011 04:01:10.027908 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-5655c58dd6-xmk8c"] Oct 11 04:01:10 crc kubenswrapper[4967]: W1011 04:01:10.037365 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podac43a6d2_5018_4d03_9a26_5da4f2619403.slice/crio-395997bef984998dbff93f3a8cbf0155992208efaf328d4c6d980b3b6c30aad1 WatchSource:0}: Error finding container 395997bef984998dbff93f3a8cbf0155992208efaf328d4c6d980b3b6c30aad1: Status 404 returned error can't find the container with id 395997bef984998dbff93f3a8cbf0155992208efaf328d4c6d980b3b6c30aad1 Oct 11 04:01:10 crc kubenswrapper[4967]: I1011 04:01:10.159176 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-7f985d654d-ncqxv"] Oct 11 04:01:10 crc kubenswrapper[4967]: W1011 04:01:10.167750 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod822cb861_f537_4669_8fc9_168f052a4b5d.slice/crio-079f734b7263f8564fdda9ad2f604913c4d5bd2b9b2948d63d42ce0c17354d05 WatchSource:0}: Error finding container 079f734b7263f8564fdda9ad2f604913c4d5bd2b9b2948d63d42ce0c17354d05: Status 404 returned error can't find the container with id 079f734b7263f8564fdda9ad2f604913c4d5bd2b9b2948d63d42ce0c17354d05 Oct 11 04:01:10 crc kubenswrapper[4967]: I1011 04:01:10.567746 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-xmk8c" event={"ID":"ac43a6d2-5018-4d03-9a26-5da4f2619403","Type":"ContainerStarted","Data":"395997bef984998dbff93f3a8cbf0155992208efaf328d4c6d980b3b6c30aad1"} Oct 11 04:01:10 crc kubenswrapper[4967]: I1011 04:01:10.569767 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-ml5f9" event={"ID":"be6fdb4b-abb9-4e76-be7e-0c7c579d3bf8","Type":"ContainerStarted","Data":"148ed647802664088e803e51f234f1092018881dc335ebaa9535fb199a45ab6e"} Oct 11 04:01:10 crc kubenswrapper[4967]: I1011 04:01:10.570448 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-ncqxv" event={"ID":"822cb861-f537-4669-8fc9-168f052a4b5d","Type":"ContainerStarted","Data":"079f734b7263f8564fdda9ad2f604913c4d5bd2b9b2948d63d42ce0c17354d05"} Oct 11 04:01:14 crc kubenswrapper[4967]: I1011 04:01:14.593060 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-5655c58dd6-xmk8c" event={"ID":"ac43a6d2-5018-4d03-9a26-5da4f2619403","Type":"ContainerStarted","Data":"de5b1e27e64e673c7a7bceb0896b30a7c386d4914682eeb85f7abc5aee2ad9b2"} Oct 11 04:01:14 crc kubenswrapper[4967]: I1011 04:01:14.594034 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-5655c58dd6-xmk8c" Oct 11 04:01:14 crc kubenswrapper[4967]: I1011 04:01:14.595322 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-5b446d88c5-ml5f9" event={"ID":"be6fdb4b-abb9-4e76-be7e-0c7c579d3bf8","Type":"ContainerStarted","Data":"953c24b4f3c6fe331e5e7420d462a8cf76f966de6105febe1279695257ba8f76"} Oct 11 04:01:14 crc kubenswrapper[4967]: I1011 04:01:14.597292 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-7f985d654d-ncqxv" event={"ID":"822cb861-f537-4669-8fc9-168f052a4b5d","Type":"ContainerStarted","Data":"22263da8462483d0741dd9a993af74f4c1d08034659597cb7b0dce4fe6f1d06f"} Oct 11 04:01:14 crc kubenswrapper[4967]: I1011 04:01:14.622339 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-5655c58dd6-xmk8c" podStartSLOduration=1.7065937469999999 podStartE2EDuration="5.622314538s" podCreationTimestamp="2025-10-11 04:01:09 +0000 UTC" firstStartedPulling="2025-10-11 04:01:10.04113219 +0000 UTC m=+598.004341113" lastFinishedPulling="2025-10-11 04:01:13.956852971 +0000 UTC m=+601.920061904" observedRunningTime="2025-10-11 04:01:14.617949232 +0000 UTC m=+602.581158165" watchObservedRunningTime="2025-10-11 04:01:14.622314538 +0000 UTC m=+602.585523671" Oct 11 04:01:14 crc kubenswrapper[4967]: I1011 04:01:14.638941 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-5b446d88c5-ml5f9" podStartSLOduration=1.688336974 podStartE2EDuration="5.638918284s" podCreationTimestamp="2025-10-11 04:01:09 +0000 UTC" firstStartedPulling="2025-10-11 04:01:09.999671687 +0000 UTC m=+597.962880620" lastFinishedPulling="2025-10-11 04:01:13.950252997 +0000 UTC m=+601.913461930" observedRunningTime="2025-10-11 04:01:14.638736719 +0000 UTC m=+602.601945652" watchObservedRunningTime="2025-10-11 04:01:14.638918284 +0000 UTC m=+602.602127217" Oct 11 04:01:14 crc kubenswrapper[4967]: I1011 04:01:14.666156 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-7f985d654d-ncqxv" podStartSLOduration=1.779257205 podStartE2EDuration="5.666130132s" podCreationTimestamp="2025-10-11 04:01:09 +0000 UTC" firstStartedPulling="2025-10-11 04:01:10.169802292 +0000 UTC m=+598.133011215" lastFinishedPulling="2025-10-11 04:01:14.056675209 +0000 UTC m=+602.019884142" observedRunningTime="2025-10-11 04:01:14.662146634 +0000 UTC m=+602.625355567" watchObservedRunningTime="2025-10-11 04:01:14.666130132 +0000 UTC m=+602.629339085" Oct 11 04:01:19 crc kubenswrapper[4967]: I1011 04:01:19.784407 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-5655c58dd6-xmk8c" Oct 11 04:01:19 crc kubenswrapper[4967]: I1011 04:01:19.964160 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-lw6dx"] Oct 11 04:01:19 crc kubenswrapper[4967]: I1011 04:01:19.964769 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="ovn-controller" containerID="cri-o://963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6" gracePeriod=30 Oct 11 04:01:19 crc kubenswrapper[4967]: I1011 04:01:19.965516 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="northd" containerID="cri-o://e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec" gracePeriod=30 Oct 11 04:01:19 crc kubenswrapper[4967]: I1011 04:01:19.965601 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="sbdb" containerID="cri-o://12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4" gracePeriod=30 Oct 11 04:01:19 crc kubenswrapper[4967]: I1011 04:01:19.965657 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="ovn-acl-logging" containerID="cri-o://37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312" gracePeriod=30 Oct 11 04:01:19 crc kubenswrapper[4967]: I1011 04:01:19.965637 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="nbdb" containerID="cri-o://20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe" gracePeriod=30 Oct 11 04:01:19 crc kubenswrapper[4967]: I1011 04:01:19.965669 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2" gracePeriod=30 Oct 11 04:01:19 crc kubenswrapper[4967]: I1011 04:01:19.965655 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="kube-rbac-proxy-node" containerID="cri-o://13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a" gracePeriod=30 Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.013824 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="ovnkube-controller" containerID="cri-o://46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1" gracePeriod=30 Oct 11 04:01:20 crc kubenswrapper[4967]: E1011 04:01:20.213516 4967 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod270ac8da_b130_4f4f_bbf7_655229b64446.slice/crio-12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4.scope\": RecentStats: unable to find data in memory cache]" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.353528 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lw6dx_270ac8da-b130-4f4f-bbf7-655229b64446/ovnkube-controller/3.log" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.361269 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lw6dx_270ac8da-b130-4f4f-bbf7-655229b64446/ovn-acl-logging/0.log" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.361979 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lw6dx_270ac8da-b130-4f4f-bbf7-655229b64446/ovn-controller/0.log" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.362787 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.436696 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-qlxsd"] Oct 11 04:01:20 crc kubenswrapper[4967]: E1011 04:01:20.437044 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="ovnkube-controller" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.437117 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="ovnkube-controller" Oct 11 04:01:20 crc kubenswrapper[4967]: E1011 04:01:20.437139 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="kube-rbac-proxy-ovn-metrics" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.437153 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="kube-rbac-proxy-ovn-metrics" Oct 11 04:01:20 crc kubenswrapper[4967]: E1011 04:01:20.437172 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="kube-rbac-proxy-node" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.437187 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="kube-rbac-proxy-node" Oct 11 04:01:20 crc kubenswrapper[4967]: E1011 04:01:20.437208 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="ovn-controller" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.437221 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="ovn-controller" Oct 11 04:01:20 crc kubenswrapper[4967]: E1011 04:01:20.437240 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="ovnkube-controller" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.437252 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="ovnkube-controller" Oct 11 04:01:20 crc kubenswrapper[4967]: E1011 04:01:20.437271 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="ovnkube-controller" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.437284 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="ovnkube-controller" Oct 11 04:01:20 crc kubenswrapper[4967]: E1011 04:01:20.437301 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="sbdb" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.437314 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="sbdb" Oct 11 04:01:20 crc kubenswrapper[4967]: E1011 04:01:20.437337 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="nbdb" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.437351 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="nbdb" Oct 11 04:01:20 crc kubenswrapper[4967]: E1011 04:01:20.437370 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="ovnkube-controller" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.437383 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="ovnkube-controller" Oct 11 04:01:20 crc kubenswrapper[4967]: E1011 04:01:20.437398 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="ovn-acl-logging" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.437410 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="ovn-acl-logging" Oct 11 04:01:20 crc kubenswrapper[4967]: E1011 04:01:20.437431 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="kubecfg-setup" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.437444 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="kubecfg-setup" Oct 11 04:01:20 crc kubenswrapper[4967]: E1011 04:01:20.437462 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="northd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.437474 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="northd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.437636 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="ovn-acl-logging" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.437657 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="kube-rbac-proxy-node" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.437679 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="kube-rbac-proxy-ovn-metrics" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.437692 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="ovnkube-controller" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.437710 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="ovnkube-controller" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.437725 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="nbdb" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.437744 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="ovn-controller" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.437758 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="sbdb" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.437775 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="ovnkube-controller" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.437792 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="northd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.437808 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="ovnkube-controller" Oct 11 04:01:20 crc kubenswrapper[4967]: E1011 04:01:20.437976 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="ovnkube-controller" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.437991 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="ovnkube-controller" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.438233 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" containerName="ovnkube-controller" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.440056 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.515451 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/270ac8da-b130-4f4f-bbf7-655229b64446-ovnkube-config\") pod \"270ac8da-b130-4f4f-bbf7-655229b64446\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.515493 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/270ac8da-b130-4f4f-bbf7-655229b64446-ovn-node-metrics-cert\") pod \"270ac8da-b130-4f4f-bbf7-655229b64446\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.515519 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-run-openvswitch\") pod \"270ac8da-b130-4f4f-bbf7-655229b64446\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.515566 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-run-ovn-kubernetes\") pod \"270ac8da-b130-4f4f-bbf7-655229b64446\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.515591 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-node-log\") pod \"270ac8da-b130-4f4f-bbf7-655229b64446\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.515612 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-etc-openvswitch\") pod \"270ac8da-b130-4f4f-bbf7-655229b64446\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.515629 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-cni-bin\") pod \"270ac8da-b130-4f4f-bbf7-655229b64446\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.515648 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/270ac8da-b130-4f4f-bbf7-655229b64446-ovnkube-script-lib\") pod \"270ac8da-b130-4f4f-bbf7-655229b64446\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.515661 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-slash\") pod \"270ac8da-b130-4f4f-bbf7-655229b64446\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.515682 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-kubelet\") pod \"270ac8da-b130-4f4f-bbf7-655229b64446\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.515700 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-log-socket\") pod \"270ac8da-b130-4f4f-bbf7-655229b64446\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.515714 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-systemd-units\") pod \"270ac8da-b130-4f4f-bbf7-655229b64446\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.515733 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8bsk\" (UniqueName: \"kubernetes.io/projected/270ac8da-b130-4f4f-bbf7-655229b64446-kube-api-access-v8bsk\") pod \"270ac8da-b130-4f4f-bbf7-655229b64446\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.515755 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-run-systemd\") pod \"270ac8da-b130-4f4f-bbf7-655229b64446\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.515774 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-var-lib-cni-networks-ovn-kubernetes\") pod \"270ac8da-b130-4f4f-bbf7-655229b64446\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.515791 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/270ac8da-b130-4f4f-bbf7-655229b64446-env-overrides\") pod \"270ac8da-b130-4f4f-bbf7-655229b64446\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.515810 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-run-ovn\") pod \"270ac8da-b130-4f4f-bbf7-655229b64446\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.515827 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-var-lib-openvswitch\") pod \"270ac8da-b130-4f4f-bbf7-655229b64446\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.515852 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-run-netns\") pod \"270ac8da-b130-4f4f-bbf7-655229b64446\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.515865 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-cni-netd\") pod \"270ac8da-b130-4f4f-bbf7-655229b64446\" (UID: \"270ac8da-b130-4f4f-bbf7-655229b64446\") " Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.516095 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "270ac8da-b130-4f4f-bbf7-655229b64446" (UID: "270ac8da-b130-4f4f-bbf7-655229b64446"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.516104 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/270ac8da-b130-4f4f-bbf7-655229b64446-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "270ac8da-b130-4f4f-bbf7-655229b64446" (UID: "270ac8da-b130-4f4f-bbf7-655229b64446"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.516182 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-log-socket" (OuterVolumeSpecName: "log-socket") pod "270ac8da-b130-4f4f-bbf7-655229b64446" (UID: "270ac8da-b130-4f4f-bbf7-655229b64446"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.516220 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "270ac8da-b130-4f4f-bbf7-655229b64446" (UID: "270ac8da-b130-4f4f-bbf7-655229b64446"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.516253 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "270ac8da-b130-4f4f-bbf7-655229b64446" (UID: "270ac8da-b130-4f4f-bbf7-655229b64446"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.516287 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-node-log" (OuterVolumeSpecName: "node-log") pod "270ac8da-b130-4f4f-bbf7-655229b64446" (UID: "270ac8da-b130-4f4f-bbf7-655229b64446"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.516317 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "270ac8da-b130-4f4f-bbf7-655229b64446" (UID: "270ac8da-b130-4f4f-bbf7-655229b64446"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.516348 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "270ac8da-b130-4f4f-bbf7-655229b64446" (UID: "270ac8da-b130-4f4f-bbf7-655229b64446"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.516740 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/270ac8da-b130-4f4f-bbf7-655229b64446-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "270ac8da-b130-4f4f-bbf7-655229b64446" (UID: "270ac8da-b130-4f4f-bbf7-655229b64446"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.516773 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "270ac8da-b130-4f4f-bbf7-655229b64446" (UID: "270ac8da-b130-4f4f-bbf7-655229b64446"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.516783 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-slash" (OuterVolumeSpecName: "host-slash") pod "270ac8da-b130-4f4f-bbf7-655229b64446" (UID: "270ac8da-b130-4f4f-bbf7-655229b64446"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.516806 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "270ac8da-b130-4f4f-bbf7-655229b64446" (UID: "270ac8da-b130-4f4f-bbf7-655229b64446"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.516830 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "270ac8da-b130-4f4f-bbf7-655229b64446" (UID: "270ac8da-b130-4f4f-bbf7-655229b64446"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.516844 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "270ac8da-b130-4f4f-bbf7-655229b64446" (UID: "270ac8da-b130-4f4f-bbf7-655229b64446"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.517466 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "270ac8da-b130-4f4f-bbf7-655229b64446" (UID: "270ac8da-b130-4f4f-bbf7-655229b64446"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.517482 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "270ac8da-b130-4f4f-bbf7-655229b64446" (UID: "270ac8da-b130-4f4f-bbf7-655229b64446"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.517594 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/270ac8da-b130-4f4f-bbf7-655229b64446-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "270ac8da-b130-4f4f-bbf7-655229b64446" (UID: "270ac8da-b130-4f4f-bbf7-655229b64446"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.521201 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/270ac8da-b130-4f4f-bbf7-655229b64446-kube-api-access-v8bsk" (OuterVolumeSpecName: "kube-api-access-v8bsk") pod "270ac8da-b130-4f4f-bbf7-655229b64446" (UID: "270ac8da-b130-4f4f-bbf7-655229b64446"). InnerVolumeSpecName "kube-api-access-v8bsk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.521424 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/270ac8da-b130-4f4f-bbf7-655229b64446-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "270ac8da-b130-4f4f-bbf7-655229b64446" (UID: "270ac8da-b130-4f4f-bbf7-655229b64446"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.539193 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "270ac8da-b130-4f4f-bbf7-655229b64446" (UID: "270ac8da-b130-4f4f-bbf7-655229b64446"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.617966 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-host-kubelet\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.618098 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-ovn-node-metrics-cert\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.618137 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-log-socket\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.618160 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-ovnkube-script-lib\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.618278 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-run-ovn\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.618302 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-host-run-ovn-kubernetes\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.618326 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-var-lib-openvswitch\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.618391 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.618420 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9rjft\" (UniqueName: \"kubernetes.io/projected/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-kube-api-access-9rjft\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.618452 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-env-overrides\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.618491 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-node-log\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.618523 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-etc-openvswitch\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.618555 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-host-slash\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.618581 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-run-systemd\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.618606 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-host-run-netns\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.618635 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-ovnkube-config\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.618659 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-host-cni-bin\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.618691 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-run-openvswitch\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.618717 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-systemd-units\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.618745 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-host-cni-netd\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.618844 4967 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-node-log\") on node \"crc\" DevicePath \"\"" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.618874 4967 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.618967 4967 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-cni-bin\") on node \"crc\" DevicePath \"\"" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.618993 4967 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/270ac8da-b130-4f4f-bbf7-655229b64446-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.619011 4967 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-slash\") on node \"crc\" DevicePath \"\"" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.619029 4967 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-kubelet\") on node \"crc\" DevicePath \"\"" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.619050 4967 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-log-socket\") on node \"crc\" DevicePath \"\"" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.619065 4967 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-systemd-units\") on node \"crc\" DevicePath \"\"" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.619112 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8bsk\" (UniqueName: \"kubernetes.io/projected/270ac8da-b130-4f4f-bbf7-655229b64446-kube-api-access-v8bsk\") on node \"crc\" DevicePath \"\"" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.619130 4967 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-run-systemd\") on node \"crc\" DevicePath \"\"" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.619148 4967 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.619163 4967 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/270ac8da-b130-4f4f-bbf7-655229b64446-env-overrides\") on node \"crc\" DevicePath \"\"" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.619177 4967 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.619191 4967 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.619207 4967 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-run-netns\") on node \"crc\" DevicePath \"\"" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.619224 4967 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-cni-netd\") on node \"crc\" DevicePath \"\"" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.619240 4967 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/270ac8da-b130-4f4f-bbf7-655229b64446-ovnkube-config\") on node \"crc\" DevicePath \"\"" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.619257 4967 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/270ac8da-b130-4f4f-bbf7-655229b64446-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.619276 4967 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-run-openvswitch\") on node \"crc\" DevicePath \"\"" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.619387 4967 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/270ac8da-b130-4f4f-bbf7-655229b64446-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.638952 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lw6dx_270ac8da-b130-4f4f-bbf7-655229b64446/ovnkube-controller/3.log" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.642625 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lw6dx_270ac8da-b130-4f4f-bbf7-655229b64446/ovn-acl-logging/0.log" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.643401 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-lw6dx_270ac8da-b130-4f4f-bbf7-655229b64446/ovn-controller/0.log" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.643974 4967 generic.go:334] "Generic (PLEG): container finished" podID="270ac8da-b130-4f4f-bbf7-655229b64446" containerID="46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1" exitCode=0 Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.644047 4967 generic.go:334] "Generic (PLEG): container finished" podID="270ac8da-b130-4f4f-bbf7-655229b64446" containerID="12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4" exitCode=0 Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.644063 4967 generic.go:334] "Generic (PLEG): container finished" podID="270ac8da-b130-4f4f-bbf7-655229b64446" containerID="20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe" exitCode=0 Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.644109 4967 generic.go:334] "Generic (PLEG): container finished" podID="270ac8da-b130-4f4f-bbf7-655229b64446" containerID="e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec" exitCode=0 Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.644123 4967 generic.go:334] "Generic (PLEG): container finished" podID="270ac8da-b130-4f4f-bbf7-655229b64446" containerID="faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2" exitCode=0 Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.644136 4967 generic.go:334] "Generic (PLEG): container finished" podID="270ac8da-b130-4f4f-bbf7-655229b64446" containerID="13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a" exitCode=0 Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.644149 4967 generic.go:334] "Generic (PLEG): container finished" podID="270ac8da-b130-4f4f-bbf7-655229b64446" containerID="37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312" exitCode=143 Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.644165 4967 generic.go:334] "Generic (PLEG): container finished" podID="270ac8da-b130-4f4f-bbf7-655229b64446" containerID="963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6" exitCode=143 Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.644189 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.645975 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" event={"ID":"270ac8da-b130-4f4f-bbf7-655229b64446","Type":"ContainerDied","Data":"46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.646221 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" event={"ID":"270ac8da-b130-4f4f-bbf7-655229b64446","Type":"ContainerDied","Data":"12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.646426 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" event={"ID":"270ac8da-b130-4f4f-bbf7-655229b64446","Type":"ContainerDied","Data":"20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.646626 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" event={"ID":"270ac8da-b130-4f4f-bbf7-655229b64446","Type":"ContainerDied","Data":"e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.646976 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" event={"ID":"270ac8da-b130-4f4f-bbf7-655229b64446","Type":"ContainerDied","Data":"faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.647238 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" event={"ID":"270ac8da-b130-4f4f-bbf7-655229b64446","Type":"ContainerDied","Data":"13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.647444 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.647622 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.647806 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.647966 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.648170 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.648430 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.648604 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.648769 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.648943 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.649185 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" event={"ID":"270ac8da-b130-4f4f-bbf7-655229b64446","Type":"ContainerDied","Data":"37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.649526 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.649726 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.649890 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.650007 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.650160 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.650327 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.650524 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.650665 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.650785 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.651019 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.651367 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" event={"ID":"270ac8da-b130-4f4f-bbf7-655229b64446","Type":"ContainerDied","Data":"963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.651655 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.651929 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.652064 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.652331 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.652670 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.652817 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.652949 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.653101 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.653249 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.653419 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.653566 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lw6dx" event={"ID":"270ac8da-b130-4f4f-bbf7-655229b64446","Type":"ContainerDied","Data":"71dca86d3ee54432a3df74b4cc1c3b83b07d1fcc602f810a6211f748cf7b3833"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.653715 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.653858 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.653981 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.654144 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.654307 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.654488 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.654626 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.654737 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.654852 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.654969 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.646338 4967 scope.go:117] "RemoveContainer" containerID="46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.647376 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6xvgx_1db525c3-0479-4585-b107-a448f2b94ffd/kube-multus/2.log" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.656451 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6xvgx_1db525c3-0479-4585-b107-a448f2b94ffd/kube-multus/1.log" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.656528 4967 generic.go:334] "Generic (PLEG): container finished" podID="1db525c3-0479-4585-b107-a448f2b94ffd" containerID="6303c6895d8a61d2d88fe615b28d6b87e1180831b83ef4f6e1019a7d186554fc" exitCode=2 Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.656568 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6xvgx" event={"ID":"1db525c3-0479-4585-b107-a448f2b94ffd","Type":"ContainerDied","Data":"6303c6895d8a61d2d88fe615b28d6b87e1180831b83ef4f6e1019a7d186554fc"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.656610 4967 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"83b1e937441999be327ba6a58c849e9c04956f5cea84da9f8f4aa476e1e79c76"} Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.657308 4967 scope.go:117] "RemoveContainer" containerID="6303c6895d8a61d2d88fe615b28d6b87e1180831b83ef4f6e1019a7d186554fc" Oct 11 04:01:20 crc kubenswrapper[4967]: E1011 04:01:20.657592 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-6xvgx_openshift-multus(1db525c3-0479-4585-b107-a448f2b94ffd)\"" pod="openshift-multus/multus-6xvgx" podUID="1db525c3-0479-4585-b107-a448f2b94ffd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.686053 4967 scope.go:117] "RemoveContainer" containerID="2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.708469 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-lw6dx"] Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.711236 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-lw6dx"] Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.720587 4967 scope.go:117] "RemoveContainer" containerID="12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.721198 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-host-kubelet\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.721254 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-ovn-node-metrics-cert\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.721284 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-log-socket\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.721308 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-ovnkube-script-lib\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.721323 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-host-kubelet\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.721353 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-run-ovn\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.721786 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-log-socket\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.721793 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-run-ovn\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.721835 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-host-run-ovn-kubernetes\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.721876 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-host-run-ovn-kubernetes\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.721921 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-var-lib-openvswitch\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.721992 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.722029 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.722002 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-var-lib-openvswitch\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.722062 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9rjft\" (UniqueName: \"kubernetes.io/projected/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-kube-api-access-9rjft\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.722171 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-env-overrides\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.722374 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-node-log\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.722392 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-ovnkube-script-lib\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.722415 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-etc-openvswitch\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.722444 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-etc-openvswitch\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.722505 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-host-slash\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.722523 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-run-systemd\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.722545 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-host-run-netns\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.722532 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-node-log\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.722560 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-ovnkube-config\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.722660 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-host-cni-bin\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.722723 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-run-openvswitch\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.722738 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-host-cni-bin\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.722759 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-systemd-units\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.722773 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-host-slash\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.722816 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-host-cni-netd\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.723015 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-host-cni-netd\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.723093 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-ovnkube-config\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.723067 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-run-openvswitch\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.723124 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-systemd-units\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.723142 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-run-systemd\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.723166 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-host-run-netns\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.723440 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-env-overrides\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.726504 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-ovn-node-metrics-cert\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.736939 4967 scope.go:117] "RemoveContainer" containerID="20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.748885 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9rjft\" (UniqueName: \"kubernetes.io/projected/f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3-kube-api-access-9rjft\") pod \"ovnkube-node-qlxsd\" (UID: \"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3\") " pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.752200 4967 scope.go:117] "RemoveContainer" containerID="e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.753362 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.771594 4967 scope.go:117] "RemoveContainer" containerID="faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2" Oct 11 04:01:20 crc kubenswrapper[4967]: W1011 04:01:20.782118 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf33ddf4c_f4e8_43ba_868f_c7f11dfb0ff3.slice/crio-6df1fb89ec3fc081c7fe7165b7dc128588b62fb2ae21806d96ba6552f5c038b1 WatchSource:0}: Error finding container 6df1fb89ec3fc081c7fe7165b7dc128588b62fb2ae21806d96ba6552f5c038b1: Status 404 returned error can't find the container with id 6df1fb89ec3fc081c7fe7165b7dc128588b62fb2ae21806d96ba6552f5c038b1 Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.789410 4967 scope.go:117] "RemoveContainer" containerID="13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.812887 4967 scope.go:117] "RemoveContainer" containerID="37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.826960 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="270ac8da-b130-4f4f-bbf7-655229b64446" path="/var/lib/kubelet/pods/270ac8da-b130-4f4f-bbf7-655229b64446/volumes" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.844584 4967 scope.go:117] "RemoveContainer" containerID="963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.860319 4967 scope.go:117] "RemoveContainer" containerID="d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.874488 4967 scope.go:117] "RemoveContainer" containerID="46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1" Oct 11 04:01:20 crc kubenswrapper[4967]: E1011 04:01:20.874907 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1\": container with ID starting with 46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1 not found: ID does not exist" containerID="46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.874972 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1"} err="failed to get container status \"46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1\": rpc error: code = NotFound desc = could not find container \"46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1\": container with ID starting with 46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1 not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.875013 4967 scope.go:117] "RemoveContainer" containerID="2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d" Oct 11 04:01:20 crc kubenswrapper[4967]: E1011 04:01:20.875497 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d\": container with ID starting with 2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d not found: ID does not exist" containerID="2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.875534 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d"} err="failed to get container status \"2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d\": rpc error: code = NotFound desc = could not find container \"2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d\": container with ID starting with 2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.875557 4967 scope.go:117] "RemoveContainer" containerID="12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4" Oct 11 04:01:20 crc kubenswrapper[4967]: E1011 04:01:20.875944 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\": container with ID starting with 12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4 not found: ID does not exist" containerID="12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.875968 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4"} err="failed to get container status \"12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\": rpc error: code = NotFound desc = could not find container \"12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\": container with ID starting with 12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4 not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.875982 4967 scope.go:117] "RemoveContainer" containerID="20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe" Oct 11 04:01:20 crc kubenswrapper[4967]: E1011 04:01:20.876430 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\": container with ID starting with 20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe not found: ID does not exist" containerID="20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.876457 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe"} err="failed to get container status \"20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\": rpc error: code = NotFound desc = could not find container \"20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\": container with ID starting with 20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.876474 4967 scope.go:117] "RemoveContainer" containerID="e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec" Oct 11 04:01:20 crc kubenswrapper[4967]: E1011 04:01:20.876739 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\": container with ID starting with e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec not found: ID does not exist" containerID="e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.876759 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec"} err="failed to get container status \"e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\": rpc error: code = NotFound desc = could not find container \"e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\": container with ID starting with e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.876771 4967 scope.go:117] "RemoveContainer" containerID="faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2" Oct 11 04:01:20 crc kubenswrapper[4967]: E1011 04:01:20.877104 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\": container with ID starting with faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2 not found: ID does not exist" containerID="faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.877141 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2"} err="failed to get container status \"faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\": rpc error: code = NotFound desc = could not find container \"faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\": container with ID starting with faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2 not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.877174 4967 scope.go:117] "RemoveContainer" containerID="13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a" Oct 11 04:01:20 crc kubenswrapper[4967]: E1011 04:01:20.877452 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\": container with ID starting with 13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a not found: ID does not exist" containerID="13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.877479 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a"} err="failed to get container status \"13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\": rpc error: code = NotFound desc = could not find container \"13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\": container with ID starting with 13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.877495 4967 scope.go:117] "RemoveContainer" containerID="37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312" Oct 11 04:01:20 crc kubenswrapper[4967]: E1011 04:01:20.877839 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\": container with ID starting with 37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312 not found: ID does not exist" containerID="37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.877873 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312"} err="failed to get container status \"37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\": rpc error: code = NotFound desc = could not find container \"37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\": container with ID starting with 37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312 not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.877894 4967 scope.go:117] "RemoveContainer" containerID="963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6" Oct 11 04:01:20 crc kubenswrapper[4967]: E1011 04:01:20.878210 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\": container with ID starting with 963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6 not found: ID does not exist" containerID="963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.878234 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6"} err="failed to get container status \"963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\": rpc error: code = NotFound desc = could not find container \"963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\": container with ID starting with 963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6 not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.878250 4967 scope.go:117] "RemoveContainer" containerID="d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1" Oct 11 04:01:20 crc kubenswrapper[4967]: E1011 04:01:20.878546 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\": container with ID starting with d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1 not found: ID does not exist" containerID="d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.878571 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1"} err="failed to get container status \"d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\": rpc error: code = NotFound desc = could not find container \"d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\": container with ID starting with d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1 not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.878586 4967 scope.go:117] "RemoveContainer" containerID="46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.878881 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1"} err="failed to get container status \"46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1\": rpc error: code = NotFound desc = could not find container \"46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1\": container with ID starting with 46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1 not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.878906 4967 scope.go:117] "RemoveContainer" containerID="2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.879171 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d"} err="failed to get container status \"2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d\": rpc error: code = NotFound desc = could not find container \"2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d\": container with ID starting with 2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.879189 4967 scope.go:117] "RemoveContainer" containerID="12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.879457 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4"} err="failed to get container status \"12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\": rpc error: code = NotFound desc = could not find container \"12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\": container with ID starting with 12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4 not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.879478 4967 scope.go:117] "RemoveContainer" containerID="20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.879717 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe"} err="failed to get container status \"20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\": rpc error: code = NotFound desc = could not find container \"20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\": container with ID starting with 20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.879734 4967 scope.go:117] "RemoveContainer" containerID="e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.880009 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec"} err="failed to get container status \"e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\": rpc error: code = NotFound desc = could not find container \"e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\": container with ID starting with e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.880035 4967 scope.go:117] "RemoveContainer" containerID="faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.880373 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2"} err="failed to get container status \"faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\": rpc error: code = NotFound desc = could not find container \"faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\": container with ID starting with faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2 not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.880401 4967 scope.go:117] "RemoveContainer" containerID="13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.880696 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a"} err="failed to get container status \"13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\": rpc error: code = NotFound desc = could not find container \"13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\": container with ID starting with 13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.880719 4967 scope.go:117] "RemoveContainer" containerID="37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.881056 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312"} err="failed to get container status \"37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\": rpc error: code = NotFound desc = could not find container \"37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\": container with ID starting with 37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312 not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.881091 4967 scope.go:117] "RemoveContainer" containerID="963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.881611 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6"} err="failed to get container status \"963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\": rpc error: code = NotFound desc = could not find container \"963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\": container with ID starting with 963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6 not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.881646 4967 scope.go:117] "RemoveContainer" containerID="d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.881952 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1"} err="failed to get container status \"d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\": rpc error: code = NotFound desc = could not find container \"d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\": container with ID starting with d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1 not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.881975 4967 scope.go:117] "RemoveContainer" containerID="46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.882357 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1"} err="failed to get container status \"46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1\": rpc error: code = NotFound desc = could not find container \"46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1\": container with ID starting with 46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1 not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.882394 4967 scope.go:117] "RemoveContainer" containerID="2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.882815 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d"} err="failed to get container status \"2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d\": rpc error: code = NotFound desc = could not find container \"2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d\": container with ID starting with 2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.882868 4967 scope.go:117] "RemoveContainer" containerID="12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.883240 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4"} err="failed to get container status \"12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\": rpc error: code = NotFound desc = could not find container \"12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\": container with ID starting with 12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4 not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.883266 4967 scope.go:117] "RemoveContainer" containerID="20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.883593 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe"} err="failed to get container status \"20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\": rpc error: code = NotFound desc = could not find container \"20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\": container with ID starting with 20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.883635 4967 scope.go:117] "RemoveContainer" containerID="e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.883960 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec"} err="failed to get container status \"e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\": rpc error: code = NotFound desc = could not find container \"e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\": container with ID starting with e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.883978 4967 scope.go:117] "RemoveContainer" containerID="faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.884430 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2"} err="failed to get container status \"faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\": rpc error: code = NotFound desc = could not find container \"faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\": container with ID starting with faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2 not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.884469 4967 scope.go:117] "RemoveContainer" containerID="13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.884813 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a"} err="failed to get container status \"13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\": rpc error: code = NotFound desc = could not find container \"13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\": container with ID starting with 13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.884856 4967 scope.go:117] "RemoveContainer" containerID="37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.885195 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312"} err="failed to get container status \"37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\": rpc error: code = NotFound desc = could not find container \"37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\": container with ID starting with 37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312 not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.885214 4967 scope.go:117] "RemoveContainer" containerID="963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.885506 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6"} err="failed to get container status \"963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\": rpc error: code = NotFound desc = could not find container \"963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\": container with ID starting with 963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6 not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.885547 4967 scope.go:117] "RemoveContainer" containerID="d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.885843 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1"} err="failed to get container status \"d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\": rpc error: code = NotFound desc = could not find container \"d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\": container with ID starting with d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1 not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.885865 4967 scope.go:117] "RemoveContainer" containerID="46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.886187 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1"} err="failed to get container status \"46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1\": rpc error: code = NotFound desc = could not find container \"46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1\": container with ID starting with 46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1 not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.886225 4967 scope.go:117] "RemoveContainer" containerID="2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.886519 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d"} err="failed to get container status \"2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d\": rpc error: code = NotFound desc = could not find container \"2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d\": container with ID starting with 2377f2128a4a7d9842584b75147925c780ef3a9bba48b9fe4e8e331d2ce0437d not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.886564 4967 scope.go:117] "RemoveContainer" containerID="12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.886942 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4"} err="failed to get container status \"12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\": rpc error: code = NotFound desc = could not find container \"12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4\": container with ID starting with 12148ac1f8e1db869ac213f6980e778f754cf845dec2a9212e57ba2bcfa3b2c4 not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.886967 4967 scope.go:117] "RemoveContainer" containerID="20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.887398 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe"} err="failed to get container status \"20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\": rpc error: code = NotFound desc = could not find container \"20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe\": container with ID starting with 20edafcaa3cc3c8c2a328d7c84558b3e0e054f7a3d569684bf238fb017dcbbfe not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.887457 4967 scope.go:117] "RemoveContainer" containerID="e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.887790 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec"} err="failed to get container status \"e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\": rpc error: code = NotFound desc = could not find container \"e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec\": container with ID starting with e8abc9a26132ae63ab3b8bf059202001fbaf3c16622aab3290d4a0c64a77edec not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.887813 4967 scope.go:117] "RemoveContainer" containerID="faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.888190 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2"} err="failed to get container status \"faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\": rpc error: code = NotFound desc = could not find container \"faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2\": container with ID starting with faf16e664d117da9768200597987f2fc8641ef64609bbb1585efe7ce375157a2 not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.888249 4967 scope.go:117] "RemoveContainer" containerID="13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.888631 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a"} err="failed to get container status \"13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\": rpc error: code = NotFound desc = could not find container \"13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a\": container with ID starting with 13e8ee9a2d2aa5f75c25ba75c229bfffff801be6a15b7f3f7948ef54b941a18a not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.888685 4967 scope.go:117] "RemoveContainer" containerID="37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.889013 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312"} err="failed to get container status \"37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\": rpc error: code = NotFound desc = could not find container \"37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312\": container with ID starting with 37a43a95e21f77b8c6aedb4636d4ea3d057348f598151c0fc117032234319312 not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.889035 4967 scope.go:117] "RemoveContainer" containerID="963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.889426 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6"} err="failed to get container status \"963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\": rpc error: code = NotFound desc = could not find container \"963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6\": container with ID starting with 963435f927385ca20aa66db07795f7a4745c1cda6d7cb7a6ee5840bb007099e6 not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.889479 4967 scope.go:117] "RemoveContainer" containerID="d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.889788 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1"} err="failed to get container status \"d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\": rpc error: code = NotFound desc = could not find container \"d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1\": container with ID starting with d7b21ac5f0f061c3770529a1c397df17155b77360e44bfc71d12a00e1e806cd1 not found: ID does not exist" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.889808 4967 scope.go:117] "RemoveContainer" containerID="46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1" Oct 11 04:01:20 crc kubenswrapper[4967]: I1011 04:01:20.890134 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1"} err="failed to get container status \"46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1\": rpc error: code = NotFound desc = could not find container \"46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1\": container with ID starting with 46c2f826322fedfa847d53d71066c32acf9af51bb06023fc8688bb56194a27a1 not found: ID does not exist" Oct 11 04:01:21 crc kubenswrapper[4967]: I1011 04:01:21.665151 4967 generic.go:334] "Generic (PLEG): container finished" podID="f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3" containerID="c613c93856c5cda30053317a8b69758652db68759c7e5195758b736a6c4d6abe" exitCode=0 Oct 11 04:01:21 crc kubenswrapper[4967]: I1011 04:01:21.665253 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" event={"ID":"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3","Type":"ContainerDied","Data":"c613c93856c5cda30053317a8b69758652db68759c7e5195758b736a6c4d6abe"} Oct 11 04:01:21 crc kubenswrapper[4967]: I1011 04:01:21.665290 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" event={"ID":"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3","Type":"ContainerStarted","Data":"6df1fb89ec3fc081c7fe7165b7dc128588b62fb2ae21806d96ba6552f5c038b1"} Oct 11 04:01:22 crc kubenswrapper[4967]: I1011 04:01:22.683246 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" event={"ID":"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3","Type":"ContainerStarted","Data":"e9902f9137e94ee7eb0534ad6361a3218fe32644002fb8d3bde7c481a0dee21f"} Oct 11 04:01:22 crc kubenswrapper[4967]: I1011 04:01:22.683545 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" event={"ID":"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3","Type":"ContainerStarted","Data":"2d4e11cf9b796aa36a96b28d4e7e6a1341fbca7e84f406938854b316e316a190"} Oct 11 04:01:22 crc kubenswrapper[4967]: I1011 04:01:22.683560 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" event={"ID":"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3","Type":"ContainerStarted","Data":"bd7037f63a2fbdf4805dcdf95829ae2b77be02fb9ec0dca0f6e2f467c69e1caf"} Oct 11 04:01:22 crc kubenswrapper[4967]: I1011 04:01:22.683574 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" event={"ID":"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3","Type":"ContainerStarted","Data":"2dd929f31df14da525b386ec35412c4cfd2d9ad9faf3e4eb6a166e00f962da94"} Oct 11 04:01:22 crc kubenswrapper[4967]: I1011 04:01:22.683587 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" event={"ID":"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3","Type":"ContainerStarted","Data":"211cf0c04238d5c284da60763934bd92d848a5dac9ff52791ef0fa1d3846316c"} Oct 11 04:01:22 crc kubenswrapper[4967]: I1011 04:01:22.683610 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" event={"ID":"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3","Type":"ContainerStarted","Data":"ad5f494a5bc81982599f495e7d62c9ddb201086104f89354817d08c51c1bb83a"} Oct 11 04:01:25 crc kubenswrapper[4967]: I1011 04:01:25.712649 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" event={"ID":"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3","Type":"ContainerStarted","Data":"d0f6a6623d606435f6437695cf3dfb609e419a5fb7d8623d2ac99956cdac5e9a"} Oct 11 04:01:27 crc kubenswrapper[4967]: I1011 04:01:27.736706 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" event={"ID":"f33ddf4c-f4e8-43ba-868f-c7f11dfb0ff3","Type":"ContainerStarted","Data":"bd9a8e257d1041603ff33c30f9e66f7b0b528ebf2e171b946f050ee91388d03c"} Oct 11 04:01:27 crc kubenswrapper[4967]: I1011 04:01:27.738257 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:27 crc kubenswrapper[4967]: I1011 04:01:27.738304 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:27 crc kubenswrapper[4967]: I1011 04:01:27.738416 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:27 crc kubenswrapper[4967]: I1011 04:01:27.779613 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:27 crc kubenswrapper[4967]: I1011 04:01:27.788038 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" podStartSLOduration=7.788002695 podStartE2EDuration="7.788002695s" podCreationTimestamp="2025-10-11 04:01:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:01:27.780688995 +0000 UTC m=+615.743897938" watchObservedRunningTime="2025-10-11 04:01:27.788002695 +0000 UTC m=+615.751211668" Oct 11 04:01:27 crc kubenswrapper[4967]: I1011 04:01:27.792196 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:31 crc kubenswrapper[4967]: I1011 04:01:31.815440 4967 scope.go:117] "RemoveContainer" containerID="6303c6895d8a61d2d88fe615b28d6b87e1180831b83ef4f6e1019a7d186554fc" Oct 11 04:01:31 crc kubenswrapper[4967]: E1011 04:01:31.817221 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-6xvgx_openshift-multus(1db525c3-0479-4585-b107-a448f2b94ffd)\"" pod="openshift-multus/multus-6xvgx" podUID="1db525c3-0479-4585-b107-a448f2b94ffd" Oct 11 04:01:42 crc kubenswrapper[4967]: I1011 04:01:42.821301 4967 scope.go:117] "RemoveContainer" containerID="6303c6895d8a61d2d88fe615b28d6b87e1180831b83ef4f6e1019a7d186554fc" Oct 11 04:01:43 crc kubenswrapper[4967]: I1011 04:01:43.845407 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6xvgx_1db525c3-0479-4585-b107-a448f2b94ffd/kube-multus/2.log" Oct 11 04:01:43 crc kubenswrapper[4967]: I1011 04:01:43.846898 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6xvgx_1db525c3-0479-4585-b107-a448f2b94ffd/kube-multus/1.log" Oct 11 04:01:43 crc kubenswrapper[4967]: I1011 04:01:43.846978 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-6xvgx" event={"ID":"1db525c3-0479-4585-b107-a448f2b94ffd","Type":"ContainerStarted","Data":"c2fcf36d0381d10facb243a58d5db1a7691a664e9429a8fbc077823cfbcd021b"} Oct 11 04:01:50 crc kubenswrapper[4967]: I1011 04:01:50.796345 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-qlxsd" Oct 11 04:01:51 crc kubenswrapper[4967]: I1011 04:01:51.837374 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6"] Oct 11 04:01:51 crc kubenswrapper[4967]: I1011 04:01:51.853704 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6" Oct 11 04:01:51 crc kubenswrapper[4967]: I1011 04:01:51.856523 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 11 04:01:51 crc kubenswrapper[4967]: I1011 04:01:51.859199 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6"] Oct 11 04:01:51 crc kubenswrapper[4967]: I1011 04:01:51.964112 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/afec4044-82ab-4b5b-8669-7605ab680e10-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6\" (UID: \"afec4044-82ab-4b5b-8669-7605ab680e10\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6" Oct 11 04:01:51 crc kubenswrapper[4967]: I1011 04:01:51.964176 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-92zpg\" (UniqueName: \"kubernetes.io/projected/afec4044-82ab-4b5b-8669-7605ab680e10-kube-api-access-92zpg\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6\" (UID: \"afec4044-82ab-4b5b-8669-7605ab680e10\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6" Oct 11 04:01:51 crc kubenswrapper[4967]: I1011 04:01:51.964991 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/afec4044-82ab-4b5b-8669-7605ab680e10-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6\" (UID: \"afec4044-82ab-4b5b-8669-7605ab680e10\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6" Oct 11 04:01:52 crc kubenswrapper[4967]: I1011 04:01:52.067039 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/afec4044-82ab-4b5b-8669-7605ab680e10-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6\" (UID: \"afec4044-82ab-4b5b-8669-7605ab680e10\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6" Oct 11 04:01:52 crc kubenswrapper[4967]: I1011 04:01:52.067152 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-92zpg\" (UniqueName: \"kubernetes.io/projected/afec4044-82ab-4b5b-8669-7605ab680e10-kube-api-access-92zpg\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6\" (UID: \"afec4044-82ab-4b5b-8669-7605ab680e10\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6" Oct 11 04:01:52 crc kubenswrapper[4967]: I1011 04:01:52.067241 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/afec4044-82ab-4b5b-8669-7605ab680e10-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6\" (UID: \"afec4044-82ab-4b5b-8669-7605ab680e10\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6" Oct 11 04:01:52 crc kubenswrapper[4967]: I1011 04:01:52.067917 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/afec4044-82ab-4b5b-8669-7605ab680e10-util\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6\" (UID: \"afec4044-82ab-4b5b-8669-7605ab680e10\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6" Oct 11 04:01:52 crc kubenswrapper[4967]: I1011 04:01:52.068129 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/afec4044-82ab-4b5b-8669-7605ab680e10-bundle\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6\" (UID: \"afec4044-82ab-4b5b-8669-7605ab680e10\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6" Oct 11 04:01:52 crc kubenswrapper[4967]: I1011 04:01:52.094176 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-92zpg\" (UniqueName: \"kubernetes.io/projected/afec4044-82ab-4b5b-8669-7605ab680e10-kube-api-access-92zpg\") pod \"fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6\" (UID: \"afec4044-82ab-4b5b-8669-7605ab680e10\") " pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6" Oct 11 04:01:52 crc kubenswrapper[4967]: I1011 04:01:52.179726 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6" Oct 11 04:01:52 crc kubenswrapper[4967]: I1011 04:01:52.404046 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6"] Oct 11 04:01:52 crc kubenswrapper[4967]: I1011 04:01:52.903704 4967 generic.go:334] "Generic (PLEG): container finished" podID="afec4044-82ab-4b5b-8669-7605ab680e10" containerID="d0da70cae0e1fc17594867d9130dcb1dab9bfad4ca6c14fbece3a9c90bc72242" exitCode=0 Oct 11 04:01:52 crc kubenswrapper[4967]: I1011 04:01:52.903849 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6" event={"ID":"afec4044-82ab-4b5b-8669-7605ab680e10","Type":"ContainerDied","Data":"d0da70cae0e1fc17594867d9130dcb1dab9bfad4ca6c14fbece3a9c90bc72242"} Oct 11 04:01:52 crc kubenswrapper[4967]: I1011 04:01:52.904064 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6" event={"ID":"afec4044-82ab-4b5b-8669-7605ab680e10","Type":"ContainerStarted","Data":"bfec2d1b2093e5f535a17a94eabaf3bf57c1796136d6e8e8094109c867921604"} Oct 11 04:01:54 crc kubenswrapper[4967]: I1011 04:01:54.919442 4967 generic.go:334] "Generic (PLEG): container finished" podID="afec4044-82ab-4b5b-8669-7605ab680e10" containerID="459b28ec8e86ed6d927a44bef925d41a747dfb2470903f70c64866bfc07bf60a" exitCode=0 Oct 11 04:01:54 crc kubenswrapper[4967]: I1011 04:01:54.919510 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6" event={"ID":"afec4044-82ab-4b5b-8669-7605ab680e10","Type":"ContainerDied","Data":"459b28ec8e86ed6d927a44bef925d41a747dfb2470903f70c64866bfc07bf60a"} Oct 11 04:01:55 crc kubenswrapper[4967]: I1011 04:01:55.927053 4967 generic.go:334] "Generic (PLEG): container finished" podID="afec4044-82ab-4b5b-8669-7605ab680e10" containerID="c19ed74fcd22c02c3cca00d124004ac5af22cf0efe0036272a266d3b55520a42" exitCode=0 Oct 11 04:01:55 crc kubenswrapper[4967]: I1011 04:01:55.927172 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6" event={"ID":"afec4044-82ab-4b5b-8669-7605ab680e10","Type":"ContainerDied","Data":"c19ed74fcd22c02c3cca00d124004ac5af22cf0efe0036272a266d3b55520a42"} Oct 11 04:01:57 crc kubenswrapper[4967]: I1011 04:01:57.275878 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6" Oct 11 04:01:57 crc kubenswrapper[4967]: I1011 04:01:57.441122 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/afec4044-82ab-4b5b-8669-7605ab680e10-util\") pod \"afec4044-82ab-4b5b-8669-7605ab680e10\" (UID: \"afec4044-82ab-4b5b-8669-7605ab680e10\") " Oct 11 04:01:57 crc kubenswrapper[4967]: I1011 04:01:57.441227 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-92zpg\" (UniqueName: \"kubernetes.io/projected/afec4044-82ab-4b5b-8669-7605ab680e10-kube-api-access-92zpg\") pod \"afec4044-82ab-4b5b-8669-7605ab680e10\" (UID: \"afec4044-82ab-4b5b-8669-7605ab680e10\") " Oct 11 04:01:57 crc kubenswrapper[4967]: I1011 04:01:57.441416 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/afec4044-82ab-4b5b-8669-7605ab680e10-bundle\") pod \"afec4044-82ab-4b5b-8669-7605ab680e10\" (UID: \"afec4044-82ab-4b5b-8669-7605ab680e10\") " Oct 11 04:01:57 crc kubenswrapper[4967]: I1011 04:01:57.442282 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afec4044-82ab-4b5b-8669-7605ab680e10-bundle" (OuterVolumeSpecName: "bundle") pod "afec4044-82ab-4b5b-8669-7605ab680e10" (UID: "afec4044-82ab-4b5b-8669-7605ab680e10"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:01:57 crc kubenswrapper[4967]: I1011 04:01:57.453411 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/afec4044-82ab-4b5b-8669-7605ab680e10-kube-api-access-92zpg" (OuterVolumeSpecName: "kube-api-access-92zpg") pod "afec4044-82ab-4b5b-8669-7605ab680e10" (UID: "afec4044-82ab-4b5b-8669-7605ab680e10"). InnerVolumeSpecName "kube-api-access-92zpg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:01:57 crc kubenswrapper[4967]: I1011 04:01:57.473195 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/afec4044-82ab-4b5b-8669-7605ab680e10-util" (OuterVolumeSpecName: "util") pod "afec4044-82ab-4b5b-8669-7605ab680e10" (UID: "afec4044-82ab-4b5b-8669-7605ab680e10"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:01:57 crc kubenswrapper[4967]: I1011 04:01:57.543244 4967 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/afec4044-82ab-4b5b-8669-7605ab680e10-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:01:57 crc kubenswrapper[4967]: I1011 04:01:57.543346 4967 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/afec4044-82ab-4b5b-8669-7605ab680e10-util\") on node \"crc\" DevicePath \"\"" Oct 11 04:01:57 crc kubenswrapper[4967]: I1011 04:01:57.543370 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-92zpg\" (UniqueName: \"kubernetes.io/projected/afec4044-82ab-4b5b-8669-7605ab680e10-kube-api-access-92zpg\") on node \"crc\" DevicePath \"\"" Oct 11 04:01:57 crc kubenswrapper[4967]: I1011 04:01:57.943276 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6" event={"ID":"afec4044-82ab-4b5b-8669-7605ab680e10","Type":"ContainerDied","Data":"bfec2d1b2093e5f535a17a94eabaf3bf57c1796136d6e8e8094109c867921604"} Oct 11 04:01:57 crc kubenswrapper[4967]: I1011 04:01:57.943342 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="bfec2d1b2093e5f535a17a94eabaf3bf57c1796136d6e8e8094109c867921604" Oct 11 04:01:57 crc kubenswrapper[4967]: I1011 04:01:57.943371 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6" Oct 11 04:01:59 crc kubenswrapper[4967]: I1011 04:01:59.339903 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-x6r8r"] Oct 11 04:01:59 crc kubenswrapper[4967]: E1011 04:01:59.340403 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afec4044-82ab-4b5b-8669-7605ab680e10" containerName="pull" Oct 11 04:01:59 crc kubenswrapper[4967]: I1011 04:01:59.340415 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="afec4044-82ab-4b5b-8669-7605ab680e10" containerName="pull" Oct 11 04:01:59 crc kubenswrapper[4967]: E1011 04:01:59.340428 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afec4044-82ab-4b5b-8669-7605ab680e10" containerName="util" Oct 11 04:01:59 crc kubenswrapper[4967]: I1011 04:01:59.340437 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="afec4044-82ab-4b5b-8669-7605ab680e10" containerName="util" Oct 11 04:01:59 crc kubenswrapper[4967]: E1011 04:01:59.340445 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="afec4044-82ab-4b5b-8669-7605ab680e10" containerName="extract" Oct 11 04:01:59 crc kubenswrapper[4967]: I1011 04:01:59.340451 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="afec4044-82ab-4b5b-8669-7605ab680e10" containerName="extract" Oct 11 04:01:59 crc kubenswrapper[4967]: I1011 04:01:59.340545 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="afec4044-82ab-4b5b-8669-7605ab680e10" containerName="extract" Oct 11 04:01:59 crc kubenswrapper[4967]: I1011 04:01:59.340910 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-x6r8r" Oct 11 04:01:59 crc kubenswrapper[4967]: I1011 04:01:59.342370 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-operator-dockercfg-ntvvk" Oct 11 04:01:59 crc kubenswrapper[4967]: I1011 04:01:59.343310 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"openshift-service-ca.crt" Oct 11 04:01:59 crc kubenswrapper[4967]: I1011 04:01:59.346831 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"kube-root-ca.crt" Oct 11 04:01:59 crc kubenswrapper[4967]: I1011 04:01:59.353256 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-x6r8r"] Oct 11 04:01:59 crc kubenswrapper[4967]: I1011 04:01:59.472686 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p472w\" (UniqueName: \"kubernetes.io/projected/419acadb-11b8-4e90-83f8-b02a39264a04-kube-api-access-p472w\") pod \"nmstate-operator-858ddd8f98-x6r8r\" (UID: \"419acadb-11b8-4e90-83f8-b02a39264a04\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-x6r8r" Oct 11 04:01:59 crc kubenswrapper[4967]: I1011 04:01:59.574591 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p472w\" (UniqueName: \"kubernetes.io/projected/419acadb-11b8-4e90-83f8-b02a39264a04-kube-api-access-p472w\") pod \"nmstate-operator-858ddd8f98-x6r8r\" (UID: \"419acadb-11b8-4e90-83f8-b02a39264a04\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-x6r8r" Oct 11 04:01:59 crc kubenswrapper[4967]: I1011 04:01:59.593168 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p472w\" (UniqueName: \"kubernetes.io/projected/419acadb-11b8-4e90-83f8-b02a39264a04-kube-api-access-p472w\") pod \"nmstate-operator-858ddd8f98-x6r8r\" (UID: \"419acadb-11b8-4e90-83f8-b02a39264a04\") " pod="openshift-nmstate/nmstate-operator-858ddd8f98-x6r8r" Oct 11 04:01:59 crc kubenswrapper[4967]: I1011 04:01:59.654910 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-operator-858ddd8f98-x6r8r" Oct 11 04:01:59 crc kubenswrapper[4967]: I1011 04:01:59.886541 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-operator-858ddd8f98-x6r8r"] Oct 11 04:01:59 crc kubenswrapper[4967]: I1011 04:01:59.956865 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-x6r8r" event={"ID":"419acadb-11b8-4e90-83f8-b02a39264a04","Type":"ContainerStarted","Data":"6212815cde17a53da856e91aaa0ad3f75c216b8821d53b383942e207a9b91ea7"} Oct 11 04:02:02 crc kubenswrapper[4967]: I1011 04:02:02.993942 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-operator-858ddd8f98-x6r8r" event={"ID":"419acadb-11b8-4e90-83f8-b02a39264a04","Type":"ContainerStarted","Data":"19121343d059c62041af31e69d6bae64f7093f345e12091c87b4355ed9e803a9"} Oct 11 04:02:03 crc kubenswrapper[4967]: I1011 04:02:03.021569 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-operator-858ddd8f98-x6r8r" podStartSLOduration=2.048811204 podStartE2EDuration="4.021549868s" podCreationTimestamp="2025-10-11 04:01:59 +0000 UTC" firstStartedPulling="2025-10-11 04:01:59.897795322 +0000 UTC m=+647.861004255" lastFinishedPulling="2025-10-11 04:02:01.870533966 +0000 UTC m=+649.833742919" observedRunningTime="2025-10-11 04:02:03.018652166 +0000 UTC m=+650.981861179" watchObservedRunningTime="2025-10-11 04:02:03.021549868 +0000 UTC m=+650.984758811" Oct 11 04:02:03 crc kubenswrapper[4967]: I1011 04:02:03.869598 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-bhrz7"] Oct 11 04:02:03 crc kubenswrapper[4967]: I1011 04:02:03.870563 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-bhrz7" Oct 11 04:02:03 crc kubenswrapper[4967]: I1011 04:02:03.873247 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"nmstate-handler-dockercfg-ppxz5" Oct 11 04:02:03 crc kubenswrapper[4967]: I1011 04:02:03.880878 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-4lvvm"] Oct 11 04:02:03 crc kubenswrapper[4967]: I1011 04:02:03.881672 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4lvvm" Oct 11 04:02:03 crc kubenswrapper[4967]: I1011 04:02:03.884734 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-bhrz7"] Oct 11 04:02:03 crc kubenswrapper[4967]: I1011 04:02:03.886666 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"openshift-nmstate-webhook" Oct 11 04:02:03 crc kubenswrapper[4967]: I1011 04:02:03.899768 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-handler-6qvqt"] Oct 11 04:02:03 crc kubenswrapper[4967]: I1011 04:02:03.900689 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-6qvqt" Oct 11 04:02:03 crc kubenswrapper[4967]: I1011 04:02:03.903967 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-4lvvm"] Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.001590 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-kfkw5"] Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.002748 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-kfkw5" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.009797 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"default-dockercfg-cp4lz" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.009992 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-nmstate"/"plugin-serving-cert" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.010151 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-nmstate"/"nginx-conf" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.017041 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-kfkw5"] Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.036195 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/b5c62540-5e0d-4601-9d0b-bef391800fe7-ovs-socket\") pod \"nmstate-handler-6qvqt\" (UID: \"b5c62540-5e0d-4601-9d0b-bef391800fe7\") " pod="openshift-nmstate/nmstate-handler-6qvqt" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.036242 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8kbt9\" (UniqueName: \"kubernetes.io/projected/abab15af-313d-42ef-96ce-4c0507429f93-kube-api-access-8kbt9\") pod \"nmstate-metrics-fdff9cb8d-bhrz7\" (UID: \"abab15af-313d-42ef-96ce-4c0507429f93\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-bhrz7" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.036263 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57jr2\" (UniqueName: \"kubernetes.io/projected/b5c62540-5e0d-4601-9d0b-bef391800fe7-kube-api-access-57jr2\") pod \"nmstate-handler-6qvqt\" (UID: \"b5c62540-5e0d-4601-9d0b-bef391800fe7\") " pod="openshift-nmstate/nmstate-handler-6qvqt" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.036294 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7bc6\" (UniqueName: \"kubernetes.io/projected/c23b1485-a6f5-4504-bc81-a069d2a96182-kube-api-access-k7bc6\") pod \"nmstate-webhook-6cdbc54649-4lvvm\" (UID: \"c23b1485-a6f5-4504-bc81-a069d2a96182\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4lvvm" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.036317 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/b5c62540-5e0d-4601-9d0b-bef391800fe7-dbus-socket\") pod \"nmstate-handler-6qvqt\" (UID: \"b5c62540-5e0d-4601-9d0b-bef391800fe7\") " pod="openshift-nmstate/nmstate-handler-6qvqt" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.036332 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/c23b1485-a6f5-4504-bc81-a069d2a96182-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-4lvvm\" (UID: \"c23b1485-a6f5-4504-bc81-a069d2a96182\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4lvvm" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.036349 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/b5c62540-5e0d-4601-9d0b-bef391800fe7-nmstate-lock\") pod \"nmstate-handler-6qvqt\" (UID: \"b5c62540-5e0d-4601-9d0b-bef391800fe7\") " pod="openshift-nmstate/nmstate-handler-6qvqt" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.137727 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/b5c62540-5e0d-4601-9d0b-bef391800fe7-ovs-socket\") pod \"nmstate-handler-6qvqt\" (UID: \"b5c62540-5e0d-4601-9d0b-bef391800fe7\") " pod="openshift-nmstate/nmstate-handler-6qvqt" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.137780 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8kbt9\" (UniqueName: \"kubernetes.io/projected/abab15af-313d-42ef-96ce-4c0507429f93-kube-api-access-8kbt9\") pod \"nmstate-metrics-fdff9cb8d-bhrz7\" (UID: \"abab15af-313d-42ef-96ce-4c0507429f93\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-bhrz7" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.137797 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57jr2\" (UniqueName: \"kubernetes.io/projected/b5c62540-5e0d-4601-9d0b-bef391800fe7-kube-api-access-57jr2\") pod \"nmstate-handler-6qvqt\" (UID: \"b5c62540-5e0d-4601-9d0b-bef391800fe7\") " pod="openshift-nmstate/nmstate-handler-6qvqt" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.137817 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/215e47d4-e289-4a93-91a7-f64b901bebeb-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-kfkw5\" (UID: \"215e47d4-e289-4a93-91a7-f64b901bebeb\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-kfkw5" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.137838 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-socket\" (UniqueName: \"kubernetes.io/host-path/b5c62540-5e0d-4601-9d0b-bef391800fe7-ovs-socket\") pod \"nmstate-handler-6qvqt\" (UID: \"b5c62540-5e0d-4601-9d0b-bef391800fe7\") " pod="openshift-nmstate/nmstate-handler-6qvqt" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.138176 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cvfck\" (UniqueName: \"kubernetes.io/projected/215e47d4-e289-4a93-91a7-f64b901bebeb-kube-api-access-cvfck\") pod \"nmstate-console-plugin-6b874cbd85-kfkw5\" (UID: \"215e47d4-e289-4a93-91a7-f64b901bebeb\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-kfkw5" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.138263 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7bc6\" (UniqueName: \"kubernetes.io/projected/c23b1485-a6f5-4504-bc81-a069d2a96182-kube-api-access-k7bc6\") pod \"nmstate-webhook-6cdbc54649-4lvvm\" (UID: \"c23b1485-a6f5-4504-bc81-a069d2a96182\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4lvvm" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.138299 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/b5c62540-5e0d-4601-9d0b-bef391800fe7-dbus-socket\") pod \"nmstate-handler-6qvqt\" (UID: \"b5c62540-5e0d-4601-9d0b-bef391800fe7\") " pod="openshift-nmstate/nmstate-handler-6qvqt" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.138336 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/c23b1485-a6f5-4504-bc81-a069d2a96182-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-4lvvm\" (UID: \"c23b1485-a6f5-4504-bc81-a069d2a96182\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4lvvm" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.138360 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/b5c62540-5e0d-4601-9d0b-bef391800fe7-nmstate-lock\") pod \"nmstate-handler-6qvqt\" (UID: \"b5c62540-5e0d-4601-9d0b-bef391800fe7\") " pod="openshift-nmstate/nmstate-handler-6qvqt" Oct 11 04:02:04 crc kubenswrapper[4967]: E1011 04:02:04.138433 4967 secret.go:188] Couldn't get secret openshift-nmstate/openshift-nmstate-webhook: secret "openshift-nmstate-webhook" not found Oct 11 04:02:04 crc kubenswrapper[4967]: E1011 04:02:04.138476 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/c23b1485-a6f5-4504-bc81-a069d2a96182-tls-key-pair podName:c23b1485-a6f5-4504-bc81-a069d2a96182 nodeName:}" failed. No retries permitted until 2025-10-11 04:02:04.638461246 +0000 UTC m=+652.601670179 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "tls-key-pair" (UniqueName: "kubernetes.io/secret/c23b1485-a6f5-4504-bc81-a069d2a96182-tls-key-pair") pod "nmstate-webhook-6cdbc54649-4lvvm" (UID: "c23b1485-a6f5-4504-bc81-a069d2a96182") : secret "openshift-nmstate-webhook" not found Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.138488 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nmstate-lock\" (UniqueName: \"kubernetes.io/host-path/b5c62540-5e0d-4601-9d0b-bef391800fe7-nmstate-lock\") pod \"nmstate-handler-6qvqt\" (UID: \"b5c62540-5e0d-4601-9d0b-bef391800fe7\") " pod="openshift-nmstate/nmstate-handler-6qvqt" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.138569 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/215e47d4-e289-4a93-91a7-f64b901bebeb-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-kfkw5\" (UID: \"215e47d4-e289-4a93-91a7-f64b901bebeb\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-kfkw5" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.138676 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dbus-socket\" (UniqueName: \"kubernetes.io/host-path/b5c62540-5e0d-4601-9d0b-bef391800fe7-dbus-socket\") pod \"nmstate-handler-6qvqt\" (UID: \"b5c62540-5e0d-4601-9d0b-bef391800fe7\") " pod="openshift-nmstate/nmstate-handler-6qvqt" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.163700 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57jr2\" (UniqueName: \"kubernetes.io/projected/b5c62540-5e0d-4601-9d0b-bef391800fe7-kube-api-access-57jr2\") pod \"nmstate-handler-6qvqt\" (UID: \"b5c62540-5e0d-4601-9d0b-bef391800fe7\") " pod="openshift-nmstate/nmstate-handler-6qvqt" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.172667 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7bc6\" (UniqueName: \"kubernetes.io/projected/c23b1485-a6f5-4504-bc81-a069d2a96182-kube-api-access-k7bc6\") pod \"nmstate-webhook-6cdbc54649-4lvvm\" (UID: \"c23b1485-a6f5-4504-bc81-a069d2a96182\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4lvvm" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.177675 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-5bbbc6d547-d9ssc"] Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.178503 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5bbbc6d547-d9ssc" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.182660 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8kbt9\" (UniqueName: \"kubernetes.io/projected/abab15af-313d-42ef-96ce-4c0507429f93-kube-api-access-8kbt9\") pod \"nmstate-metrics-fdff9cb8d-bhrz7\" (UID: \"abab15af-313d-42ef-96ce-4c0507429f93\") " pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-bhrz7" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.184276 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-bhrz7" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.198855 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5bbbc6d547-d9ssc"] Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.218335 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-handler-6qvqt" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.242463 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/215e47d4-e289-4a93-91a7-f64b901bebeb-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-kfkw5\" (UID: \"215e47d4-e289-4a93-91a7-f64b901bebeb\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-kfkw5" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.242540 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/215e47d4-e289-4a93-91a7-f64b901bebeb-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-kfkw5\" (UID: \"215e47d4-e289-4a93-91a7-f64b901bebeb\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-kfkw5" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.242565 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cvfck\" (UniqueName: \"kubernetes.io/projected/215e47d4-e289-4a93-91a7-f64b901bebeb-kube-api-access-cvfck\") pod \"nmstate-console-plugin-6b874cbd85-kfkw5\" (UID: \"215e47d4-e289-4a93-91a7-f64b901bebeb\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-kfkw5" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.244460 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/215e47d4-e289-4a93-91a7-f64b901bebeb-nginx-conf\") pod \"nmstate-console-plugin-6b874cbd85-kfkw5\" (UID: \"215e47d4-e289-4a93-91a7-f64b901bebeb\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-kfkw5" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.247882 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugin-serving-cert\" (UniqueName: \"kubernetes.io/secret/215e47d4-e289-4a93-91a7-f64b901bebeb-plugin-serving-cert\") pod \"nmstate-console-plugin-6b874cbd85-kfkw5\" (UID: \"215e47d4-e289-4a93-91a7-f64b901bebeb\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-kfkw5" Oct 11 04:02:04 crc kubenswrapper[4967]: W1011 04:02:04.251206 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb5c62540_5e0d_4601_9d0b_bef391800fe7.slice/crio-b6c982c6cc6db3841dc8406e20e54bb65a3cd4a2c5995129ef5f580034e04c16 WatchSource:0}: Error finding container b6c982c6cc6db3841dc8406e20e54bb65a3cd4a2c5995129ef5f580034e04c16: Status 404 returned error can't find the container with id b6c982c6cc6db3841dc8406e20e54bb65a3cd4a2c5995129ef5f580034e04c16 Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.261461 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cvfck\" (UniqueName: \"kubernetes.io/projected/215e47d4-e289-4a93-91a7-f64b901bebeb-kube-api-access-cvfck\") pod \"nmstate-console-plugin-6b874cbd85-kfkw5\" (UID: \"215e47d4-e289-4a93-91a7-f64b901bebeb\") " pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-kfkw5" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.322457 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-kfkw5" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.343346 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/81433b9c-bcea-468c-b15c-3938d1a021d3-console-serving-cert\") pod \"console-5bbbc6d547-d9ssc\" (UID: \"81433b9c-bcea-468c-b15c-3938d1a021d3\") " pod="openshift-console/console-5bbbc6d547-d9ssc" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.343416 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/81433b9c-bcea-468c-b15c-3938d1a021d3-console-oauth-config\") pod \"console-5bbbc6d547-d9ssc\" (UID: \"81433b9c-bcea-468c-b15c-3938d1a021d3\") " pod="openshift-console/console-5bbbc6d547-d9ssc" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.343433 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/81433b9c-bcea-468c-b15c-3938d1a021d3-service-ca\") pod \"console-5bbbc6d547-d9ssc\" (UID: \"81433b9c-bcea-468c-b15c-3938d1a021d3\") " pod="openshift-console/console-5bbbc6d547-d9ssc" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.343453 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/81433b9c-bcea-468c-b15c-3938d1a021d3-oauth-serving-cert\") pod \"console-5bbbc6d547-d9ssc\" (UID: \"81433b9c-bcea-468c-b15c-3938d1a021d3\") " pod="openshift-console/console-5bbbc6d547-d9ssc" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.343493 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/81433b9c-bcea-468c-b15c-3938d1a021d3-trusted-ca-bundle\") pod \"console-5bbbc6d547-d9ssc\" (UID: \"81433b9c-bcea-468c-b15c-3938d1a021d3\") " pod="openshift-console/console-5bbbc6d547-d9ssc" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.343513 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9qsj8\" (UniqueName: \"kubernetes.io/projected/81433b9c-bcea-468c-b15c-3938d1a021d3-kube-api-access-9qsj8\") pod \"console-5bbbc6d547-d9ssc\" (UID: \"81433b9c-bcea-468c-b15c-3938d1a021d3\") " pod="openshift-console/console-5bbbc6d547-d9ssc" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.344619 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/81433b9c-bcea-468c-b15c-3938d1a021d3-console-config\") pod \"console-5bbbc6d547-d9ssc\" (UID: \"81433b9c-bcea-468c-b15c-3938d1a021d3\") " pod="openshift-console/console-5bbbc6d547-d9ssc" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.412599 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-metrics-fdff9cb8d-bhrz7"] Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.445498 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/81433b9c-bcea-468c-b15c-3938d1a021d3-console-serving-cert\") pod \"console-5bbbc6d547-d9ssc\" (UID: \"81433b9c-bcea-468c-b15c-3938d1a021d3\") " pod="openshift-console/console-5bbbc6d547-d9ssc" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.445537 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/81433b9c-bcea-468c-b15c-3938d1a021d3-console-oauth-config\") pod \"console-5bbbc6d547-d9ssc\" (UID: \"81433b9c-bcea-468c-b15c-3938d1a021d3\") " pod="openshift-console/console-5bbbc6d547-d9ssc" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.445593 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/81433b9c-bcea-468c-b15c-3938d1a021d3-service-ca\") pod \"console-5bbbc6d547-d9ssc\" (UID: \"81433b9c-bcea-468c-b15c-3938d1a021d3\") " pod="openshift-console/console-5bbbc6d547-d9ssc" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.445617 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/81433b9c-bcea-468c-b15c-3938d1a021d3-oauth-serving-cert\") pod \"console-5bbbc6d547-d9ssc\" (UID: \"81433b9c-bcea-468c-b15c-3938d1a021d3\") " pod="openshift-console/console-5bbbc6d547-d9ssc" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.445657 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/81433b9c-bcea-468c-b15c-3938d1a021d3-trusted-ca-bundle\") pod \"console-5bbbc6d547-d9ssc\" (UID: \"81433b9c-bcea-468c-b15c-3938d1a021d3\") " pod="openshift-console/console-5bbbc6d547-d9ssc" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.445681 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9qsj8\" (UniqueName: \"kubernetes.io/projected/81433b9c-bcea-468c-b15c-3938d1a021d3-kube-api-access-9qsj8\") pod \"console-5bbbc6d547-d9ssc\" (UID: \"81433b9c-bcea-468c-b15c-3938d1a021d3\") " pod="openshift-console/console-5bbbc6d547-d9ssc" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.445711 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/81433b9c-bcea-468c-b15c-3938d1a021d3-console-config\") pod \"console-5bbbc6d547-d9ssc\" (UID: \"81433b9c-bcea-468c-b15c-3938d1a021d3\") " pod="openshift-console/console-5bbbc6d547-d9ssc" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.447190 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/81433b9c-bcea-468c-b15c-3938d1a021d3-console-config\") pod \"console-5bbbc6d547-d9ssc\" (UID: \"81433b9c-bcea-468c-b15c-3938d1a021d3\") " pod="openshift-console/console-5bbbc6d547-d9ssc" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.448360 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/81433b9c-bcea-468c-b15c-3938d1a021d3-oauth-serving-cert\") pod \"console-5bbbc6d547-d9ssc\" (UID: \"81433b9c-bcea-468c-b15c-3938d1a021d3\") " pod="openshift-console/console-5bbbc6d547-d9ssc" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.450184 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/81433b9c-bcea-468c-b15c-3938d1a021d3-console-serving-cert\") pod \"console-5bbbc6d547-d9ssc\" (UID: \"81433b9c-bcea-468c-b15c-3938d1a021d3\") " pod="openshift-console/console-5bbbc6d547-d9ssc" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.450826 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/81433b9c-bcea-468c-b15c-3938d1a021d3-service-ca\") pod \"console-5bbbc6d547-d9ssc\" (UID: \"81433b9c-bcea-468c-b15c-3938d1a021d3\") " pod="openshift-console/console-5bbbc6d547-d9ssc" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.451027 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/81433b9c-bcea-468c-b15c-3938d1a021d3-trusted-ca-bundle\") pod \"console-5bbbc6d547-d9ssc\" (UID: \"81433b9c-bcea-468c-b15c-3938d1a021d3\") " pod="openshift-console/console-5bbbc6d547-d9ssc" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.451448 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/81433b9c-bcea-468c-b15c-3938d1a021d3-console-oauth-config\") pod \"console-5bbbc6d547-d9ssc\" (UID: \"81433b9c-bcea-468c-b15c-3938d1a021d3\") " pod="openshift-console/console-5bbbc6d547-d9ssc" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.464203 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9qsj8\" (UniqueName: \"kubernetes.io/projected/81433b9c-bcea-468c-b15c-3938d1a021d3-kube-api-access-9qsj8\") pod \"console-5bbbc6d547-d9ssc\" (UID: \"81433b9c-bcea-468c-b15c-3938d1a021d3\") " pod="openshift-console/console-5bbbc6d547-d9ssc" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.542595 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-5bbbc6d547-d9ssc" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.648144 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/c23b1485-a6f5-4504-bc81-a069d2a96182-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-4lvvm\" (UID: \"c23b1485-a6f5-4504-bc81-a069d2a96182\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4lvvm" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.651343 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tls-key-pair\" (UniqueName: \"kubernetes.io/secret/c23b1485-a6f5-4504-bc81-a069d2a96182-tls-key-pair\") pod \"nmstate-webhook-6cdbc54649-4lvvm\" (UID: \"c23b1485-a6f5-4504-bc81-a069d2a96182\") " pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4lvvm" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.713195 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-5bbbc6d547-d9ssc"] Oct 11 04:02:04 crc kubenswrapper[4967]: W1011 04:02:04.716546 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod81433b9c_bcea_468c_b15c_3938d1a021d3.slice/crio-22903c3ccb9298e9ec9c79c153f683d8f8e4b240f98dbf1be18c6adabcfe4d4a WatchSource:0}: Error finding container 22903c3ccb9298e9ec9c79c153f683d8f8e4b240f98dbf1be18c6adabcfe4d4a: Status 404 returned error can't find the container with id 22903c3ccb9298e9ec9c79c153f683d8f8e4b240f98dbf1be18c6adabcfe4d4a Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.766460 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-console-plugin-6b874cbd85-kfkw5"] Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.797428 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4lvvm" Oct 11 04:02:04 crc kubenswrapper[4967]: I1011 04:02:04.955816 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-nmstate/nmstate-webhook-6cdbc54649-4lvvm"] Oct 11 04:02:04 crc kubenswrapper[4967]: W1011 04:02:04.963183 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc23b1485_a6f5_4504_bc81_a069d2a96182.slice/crio-6e8ddbfa687537a1e76e988b09fff252bc1a6931366785f34b2936bba093ab11 WatchSource:0}: Error finding container 6e8ddbfa687537a1e76e988b09fff252bc1a6931366785f34b2936bba093ab11: Status 404 returned error can't find the container with id 6e8ddbfa687537a1e76e988b09fff252bc1a6931366785f34b2936bba093ab11 Oct 11 04:02:05 crc kubenswrapper[4967]: I1011 04:02:05.010118 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-6qvqt" event={"ID":"b5c62540-5e0d-4601-9d0b-bef391800fe7","Type":"ContainerStarted","Data":"b6c982c6cc6db3841dc8406e20e54bb65a3cd4a2c5995129ef5f580034e04c16"} Oct 11 04:02:05 crc kubenswrapper[4967]: I1011 04:02:05.011220 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4lvvm" event={"ID":"c23b1485-a6f5-4504-bc81-a069d2a96182","Type":"ContainerStarted","Data":"6e8ddbfa687537a1e76e988b09fff252bc1a6931366785f34b2936bba093ab11"} Oct 11 04:02:05 crc kubenswrapper[4967]: I1011 04:02:05.012602 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5bbbc6d547-d9ssc" event={"ID":"81433b9c-bcea-468c-b15c-3938d1a021d3","Type":"ContainerStarted","Data":"2ead027870bb29152b8668c5415f3bc5a8b0dd41d1c1f0322849720825393010"} Oct 11 04:02:05 crc kubenswrapper[4967]: I1011 04:02:05.012625 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-5bbbc6d547-d9ssc" event={"ID":"81433b9c-bcea-468c-b15c-3938d1a021d3","Type":"ContainerStarted","Data":"22903c3ccb9298e9ec9c79c153f683d8f8e4b240f98dbf1be18c6adabcfe4d4a"} Oct 11 04:02:05 crc kubenswrapper[4967]: I1011 04:02:05.014698 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-kfkw5" event={"ID":"215e47d4-e289-4a93-91a7-f64b901bebeb","Type":"ContainerStarted","Data":"ceeb83397aeb5a009c490db8b7483e02aa6596e670f79dc177fdc7a295399322"} Oct 11 04:02:05 crc kubenswrapper[4967]: I1011 04:02:05.016066 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-bhrz7" event={"ID":"abab15af-313d-42ef-96ce-4c0507429f93","Type":"ContainerStarted","Data":"ccdd56db535ff376d78645ff0fe3d9219eeee0a5effa28abfef0bc86759ae3b1"} Oct 11 04:02:07 crc kubenswrapper[4967]: I1011 04:02:07.031198 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4lvvm" event={"ID":"c23b1485-a6f5-4504-bc81-a069d2a96182","Type":"ContainerStarted","Data":"2ff79755a6889773b32685d1dd74f8e228c187cc22602ecc9f6993c6890430b2"} Oct 11 04:02:07 crc kubenswrapper[4967]: I1011 04:02:07.031543 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4lvvm" Oct 11 04:02:07 crc kubenswrapper[4967]: I1011 04:02:07.034769 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-bhrz7" event={"ID":"abab15af-313d-42ef-96ce-4c0507429f93","Type":"ContainerStarted","Data":"133e5f133dd9b3e704420cec3868182d6e8a52d07d73f385d0e8a6da56b0c5f9"} Oct 11 04:02:07 crc kubenswrapper[4967]: I1011 04:02:07.036255 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-handler-6qvqt" event={"ID":"b5c62540-5e0d-4601-9d0b-bef391800fe7","Type":"ContainerStarted","Data":"122e1e6c88b4b86d0cfee91724728484f4d156432119f5e62330a052018b16fb"} Oct 11 04:02:07 crc kubenswrapper[4967]: I1011 04:02:07.036688 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-nmstate/nmstate-handler-6qvqt" Oct 11 04:02:07 crc kubenswrapper[4967]: I1011 04:02:07.049004 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-5bbbc6d547-d9ssc" podStartSLOduration=3.048981799 podStartE2EDuration="3.048981799s" podCreationTimestamp="2025-10-11 04:02:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:02:05.031305851 +0000 UTC m=+652.994514784" watchObservedRunningTime="2025-10-11 04:02:07.048981799 +0000 UTC m=+655.012190752" Oct 11 04:02:07 crc kubenswrapper[4967]: I1011 04:02:07.052773 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4lvvm" podStartSLOduration=2.363791488 podStartE2EDuration="4.052757892s" podCreationTimestamp="2025-10-11 04:02:03 +0000 UTC" firstStartedPulling="2025-10-11 04:02:04.965855351 +0000 UTC m=+652.929064284" lastFinishedPulling="2025-10-11 04:02:06.654821755 +0000 UTC m=+654.618030688" observedRunningTime="2025-10-11 04:02:07.048819775 +0000 UTC m=+655.012028708" watchObservedRunningTime="2025-10-11 04:02:07.052757892 +0000 UTC m=+655.015966835" Oct 11 04:02:07 crc kubenswrapper[4967]: I1011 04:02:07.076035 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-handler-6qvqt" podStartSLOduration=1.715394253 podStartE2EDuration="4.076011488s" podCreationTimestamp="2025-10-11 04:02:03 +0000 UTC" firstStartedPulling="2025-10-11 04:02:04.273852147 +0000 UTC m=+652.237061080" lastFinishedPulling="2025-10-11 04:02:06.634469382 +0000 UTC m=+654.597678315" observedRunningTime="2025-10-11 04:02:07.066356559 +0000 UTC m=+655.029565512" watchObservedRunningTime="2025-10-11 04:02:07.076011488 +0000 UTC m=+655.039220431" Oct 11 04:02:08 crc kubenswrapper[4967]: I1011 04:02:08.048260 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-kfkw5" event={"ID":"215e47d4-e289-4a93-91a7-f64b901bebeb","Type":"ContainerStarted","Data":"c2f8dfe0c9ed3a80a404a4766eb05a1cf60d4ff54951d5387d7ba8e4088c2557"} Oct 11 04:02:08 crc kubenswrapper[4967]: I1011 04:02:08.080390 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-console-plugin-6b874cbd85-kfkw5" podStartSLOduration=2.294503032 podStartE2EDuration="5.0803608s" podCreationTimestamp="2025-10-11 04:02:03 +0000 UTC" firstStartedPulling="2025-10-11 04:02:04.777186412 +0000 UTC m=+652.740395345" lastFinishedPulling="2025-10-11 04:02:07.56304418 +0000 UTC m=+655.526253113" observedRunningTime="2025-10-11 04:02:08.072746042 +0000 UTC m=+656.035955035" watchObservedRunningTime="2025-10-11 04:02:08.0803608 +0000 UTC m=+656.043569773" Oct 11 04:02:10 crc kubenswrapper[4967]: I1011 04:02:10.071749 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-bhrz7" event={"ID":"abab15af-313d-42ef-96ce-4c0507429f93","Type":"ContainerStarted","Data":"face3d982b903faaeb4a74aa600e31461ac7db1a8fe3a521d79d17f9b8274310"} Oct 11 04:02:10 crc kubenswrapper[4967]: I1011 04:02:10.098273 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-nmstate/nmstate-metrics-fdff9cb8d-bhrz7" podStartSLOduration=2.028743707 podStartE2EDuration="7.098244334s" podCreationTimestamp="2025-10-11 04:02:03 +0000 UTC" firstStartedPulling="2025-10-11 04:02:04.472324528 +0000 UTC m=+652.435533461" lastFinishedPulling="2025-10-11 04:02:09.541825125 +0000 UTC m=+657.505034088" observedRunningTime="2025-10-11 04:02:10.092518702 +0000 UTC m=+658.055727635" watchObservedRunningTime="2025-10-11 04:02:10.098244334 +0000 UTC m=+658.061453317" Oct 11 04:02:13 crc kubenswrapper[4967]: I1011 04:02:13.070861 4967 scope.go:117] "RemoveContainer" containerID="83b1e937441999be327ba6a58c849e9c04956f5cea84da9f8f4aa476e1e79c76" Oct 11 04:02:14 crc kubenswrapper[4967]: I1011 04:02:14.105423 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-6xvgx_1db525c3-0479-4585-b107-a448f2b94ffd/kube-multus/2.log" Oct 11 04:02:14 crc kubenswrapper[4967]: I1011 04:02:14.243416 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-handler-6qvqt" Oct 11 04:02:14 crc kubenswrapper[4967]: I1011 04:02:14.543483 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-5bbbc6d547-d9ssc" Oct 11 04:02:14 crc kubenswrapper[4967]: I1011 04:02:14.544408 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-5bbbc6d547-d9ssc" Oct 11 04:02:14 crc kubenswrapper[4967]: I1011 04:02:14.556435 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-5bbbc6d547-d9ssc" Oct 11 04:02:15 crc kubenswrapper[4967]: I1011 04:02:15.115741 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-5bbbc6d547-d9ssc" Oct 11 04:02:15 crc kubenswrapper[4967]: I1011 04:02:15.205771 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-csl6r"] Oct 11 04:02:24 crc kubenswrapper[4967]: I1011 04:02:24.807182 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-nmstate/nmstate-webhook-6cdbc54649-4lvvm" Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.244549 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn"] Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.249727 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn" Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.250257 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-console/console-f9d7485db-csl6r" podUID="5e20a512-8f34-4d21-be45-e749f7a6cc87" containerName="console" containerID="cri-o://e8cf8cba0e027047127e72821d4240c617453de6ea5d373ed64d907303c53a6e" gracePeriod=15 Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.257613 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.262050 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn"] Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.365776 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6fdg6\" (UniqueName: \"kubernetes.io/projected/bd662d9c-6ab0-4518-8fab-83115ccd3569-kube-api-access-6fdg6\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn\" (UID: \"bd662d9c-6ab0-4518-8fab-83115ccd3569\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn" Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.365818 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd662d9c-6ab0-4518-8fab-83115ccd3569-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn\" (UID: \"bd662d9c-6ab0-4518-8fab-83115ccd3569\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn" Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.365877 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd662d9c-6ab0-4518-8fab-83115ccd3569-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn\" (UID: \"bd662d9c-6ab0-4518-8fab-83115ccd3569\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn" Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.467514 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6fdg6\" (UniqueName: \"kubernetes.io/projected/bd662d9c-6ab0-4518-8fab-83115ccd3569-kube-api-access-6fdg6\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn\" (UID: \"bd662d9c-6ab0-4518-8fab-83115ccd3569\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn" Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.467555 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd662d9c-6ab0-4518-8fab-83115ccd3569-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn\" (UID: \"bd662d9c-6ab0-4518-8fab-83115ccd3569\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn" Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.467614 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd662d9c-6ab0-4518-8fab-83115ccd3569-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn\" (UID: \"bd662d9c-6ab0-4518-8fab-83115ccd3569\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn" Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.468129 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd662d9c-6ab0-4518-8fab-83115ccd3569-bundle\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn\" (UID: \"bd662d9c-6ab0-4518-8fab-83115ccd3569\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn" Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.468325 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd662d9c-6ab0-4518-8fab-83115ccd3569-util\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn\" (UID: \"bd662d9c-6ab0-4518-8fab-83115ccd3569\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn" Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.505610 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6fdg6\" (UniqueName: \"kubernetes.io/projected/bd662d9c-6ab0-4518-8fab-83115ccd3569-kube-api-access-6fdg6\") pod \"8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn\" (UID: \"bd662d9c-6ab0-4518-8fab-83115ccd3569\") " pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn" Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.601487 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-csl6r_5e20a512-8f34-4d21-be45-e749f7a6cc87/console/0.log" Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.601576 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-csl6r" Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.631923 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn" Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.771444 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5e20a512-8f34-4d21-be45-e749f7a6cc87-console-config\") pod \"5e20a512-8f34-4d21-be45-e749f7a6cc87\" (UID: \"5e20a512-8f34-4d21-be45-e749f7a6cc87\") " Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.771540 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5e20a512-8f34-4d21-be45-e749f7a6cc87-console-oauth-config\") pod \"5e20a512-8f34-4d21-be45-e749f7a6cc87\" (UID: \"5e20a512-8f34-4d21-be45-e749f7a6cc87\") " Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.771794 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zwlq2\" (UniqueName: \"kubernetes.io/projected/5e20a512-8f34-4d21-be45-e749f7a6cc87-kube-api-access-zwlq2\") pod \"5e20a512-8f34-4d21-be45-e749f7a6cc87\" (UID: \"5e20a512-8f34-4d21-be45-e749f7a6cc87\") " Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.771942 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5e20a512-8f34-4d21-be45-e749f7a6cc87-service-ca\") pod \"5e20a512-8f34-4d21-be45-e749f7a6cc87\" (UID: \"5e20a512-8f34-4d21-be45-e749f7a6cc87\") " Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.772031 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e20a512-8f34-4d21-be45-e749f7a6cc87-trusted-ca-bundle\") pod \"5e20a512-8f34-4d21-be45-e749f7a6cc87\" (UID: \"5e20a512-8f34-4d21-be45-e749f7a6cc87\") " Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.772179 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5e20a512-8f34-4d21-be45-e749f7a6cc87-oauth-serving-cert\") pod \"5e20a512-8f34-4d21-be45-e749f7a6cc87\" (UID: \"5e20a512-8f34-4d21-be45-e749f7a6cc87\") " Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.772422 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5e20a512-8f34-4d21-be45-e749f7a6cc87-console-serving-cert\") pod \"5e20a512-8f34-4d21-be45-e749f7a6cc87\" (UID: \"5e20a512-8f34-4d21-be45-e749f7a6cc87\") " Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.776616 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e20a512-8f34-4d21-be45-e749f7a6cc87-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "5e20a512-8f34-4d21-be45-e749f7a6cc87" (UID: "5e20a512-8f34-4d21-be45-e749f7a6cc87"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.776797 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e20a512-8f34-4d21-be45-e749f7a6cc87-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "5e20a512-8f34-4d21-be45-e749f7a6cc87" (UID: "5e20a512-8f34-4d21-be45-e749f7a6cc87"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.777220 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e20a512-8f34-4d21-be45-e749f7a6cc87-service-ca" (OuterVolumeSpecName: "service-ca") pod "5e20a512-8f34-4d21-be45-e749f7a6cc87" (UID: "5e20a512-8f34-4d21-be45-e749f7a6cc87"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.777563 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e20a512-8f34-4d21-be45-e749f7a6cc87-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "5e20a512-8f34-4d21-be45-e749f7a6cc87" (UID: "5e20a512-8f34-4d21-be45-e749f7a6cc87"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.777579 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e20a512-8f34-4d21-be45-e749f7a6cc87-console-config" (OuterVolumeSpecName: "console-config") pod "5e20a512-8f34-4d21-be45-e749f7a6cc87" (UID: "5e20a512-8f34-4d21-be45-e749f7a6cc87"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.781460 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e20a512-8f34-4d21-be45-e749f7a6cc87-kube-api-access-zwlq2" (OuterVolumeSpecName: "kube-api-access-zwlq2") pod "5e20a512-8f34-4d21-be45-e749f7a6cc87" (UID: "5e20a512-8f34-4d21-be45-e749f7a6cc87"). InnerVolumeSpecName "kube-api-access-zwlq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.783845 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e20a512-8f34-4d21-be45-e749f7a6cc87-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "5e20a512-8f34-4d21-be45-e749f7a6cc87" (UID: "5e20a512-8f34-4d21-be45-e749f7a6cc87"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.810577 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn"] Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.874039 4967 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/5e20a512-8f34-4d21-be45-e749f7a6cc87-console-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.874074 4967 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/5e20a512-8f34-4d21-be45-e749f7a6cc87-console-config\") on node \"crc\" DevicePath \"\"" Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.874105 4967 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/5e20a512-8f34-4d21-be45-e749f7a6cc87-console-oauth-config\") on node \"crc\" DevicePath \"\"" Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.874113 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zwlq2\" (UniqueName: \"kubernetes.io/projected/5e20a512-8f34-4d21-be45-e749f7a6cc87-kube-api-access-zwlq2\") on node \"crc\" DevicePath \"\"" Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.874122 4967 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/5e20a512-8f34-4d21-be45-e749f7a6cc87-service-ca\") on node \"crc\" DevicePath \"\"" Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.874130 4967 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5e20a512-8f34-4d21-be45-e749f7a6cc87-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:02:40 crc kubenswrapper[4967]: I1011 04:02:40.874138 4967 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/5e20a512-8f34-4d21-be45-e749f7a6cc87-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 04:02:41 crc kubenswrapper[4967]: I1011 04:02:41.305558 4967 generic.go:334] "Generic (PLEG): container finished" podID="bd662d9c-6ab0-4518-8fab-83115ccd3569" containerID="b17fea74422c82a69df7ddfd3d49ea3b85d537d0799f9d200ff6af8077b45e51" exitCode=0 Oct 11 04:02:41 crc kubenswrapper[4967]: I1011 04:02:41.305692 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn" event={"ID":"bd662d9c-6ab0-4518-8fab-83115ccd3569","Type":"ContainerDied","Data":"b17fea74422c82a69df7ddfd3d49ea3b85d537d0799f9d200ff6af8077b45e51"} Oct 11 04:02:41 crc kubenswrapper[4967]: I1011 04:02:41.307516 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn" event={"ID":"bd662d9c-6ab0-4518-8fab-83115ccd3569","Type":"ContainerStarted","Data":"cee8826d98d53ea127cfad2a704c28c9a60d98d4d2fd234e11b0afa44ee1bda9"} Oct 11 04:02:41 crc kubenswrapper[4967]: I1011 04:02:41.313479 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-console_console-f9d7485db-csl6r_5e20a512-8f34-4d21-be45-e749f7a6cc87/console/0.log" Oct 11 04:02:41 crc kubenswrapper[4967]: I1011 04:02:41.313551 4967 generic.go:334] "Generic (PLEG): container finished" podID="5e20a512-8f34-4d21-be45-e749f7a6cc87" containerID="e8cf8cba0e027047127e72821d4240c617453de6ea5d373ed64d907303c53a6e" exitCode=2 Oct 11 04:02:41 crc kubenswrapper[4967]: I1011 04:02:41.313594 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-csl6r" event={"ID":"5e20a512-8f34-4d21-be45-e749f7a6cc87","Type":"ContainerDied","Data":"e8cf8cba0e027047127e72821d4240c617453de6ea5d373ed64d907303c53a6e"} Oct 11 04:02:41 crc kubenswrapper[4967]: I1011 04:02:41.313633 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-csl6r" event={"ID":"5e20a512-8f34-4d21-be45-e749f7a6cc87","Type":"ContainerDied","Data":"34766231a2f79dc50e67406c04d6e72ed0012f85c3c3e3ca47d3ff1e82eaa613"} Oct 11 04:02:41 crc kubenswrapper[4967]: I1011 04:02:41.313661 4967 scope.go:117] "RemoveContainer" containerID="e8cf8cba0e027047127e72821d4240c617453de6ea5d373ed64d907303c53a6e" Oct 11 04:02:41 crc kubenswrapper[4967]: I1011 04:02:41.313687 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-csl6r" Oct 11 04:02:41 crc kubenswrapper[4967]: I1011 04:02:41.356316 4967 scope.go:117] "RemoveContainer" containerID="e8cf8cba0e027047127e72821d4240c617453de6ea5d373ed64d907303c53a6e" Oct 11 04:02:41 crc kubenswrapper[4967]: I1011 04:02:41.359485 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-console/console-f9d7485db-csl6r"] Oct 11 04:02:41 crc kubenswrapper[4967]: E1011 04:02:41.362406 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e8cf8cba0e027047127e72821d4240c617453de6ea5d373ed64d907303c53a6e\": container with ID starting with e8cf8cba0e027047127e72821d4240c617453de6ea5d373ed64d907303c53a6e not found: ID does not exist" containerID="e8cf8cba0e027047127e72821d4240c617453de6ea5d373ed64d907303c53a6e" Oct 11 04:02:41 crc kubenswrapper[4967]: I1011 04:02:41.362482 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e8cf8cba0e027047127e72821d4240c617453de6ea5d373ed64d907303c53a6e"} err="failed to get container status \"e8cf8cba0e027047127e72821d4240c617453de6ea5d373ed64d907303c53a6e\": rpc error: code = NotFound desc = could not find container \"e8cf8cba0e027047127e72821d4240c617453de6ea5d373ed64d907303c53a6e\": container with ID starting with e8cf8cba0e027047127e72821d4240c617453de6ea5d373ed64d907303c53a6e not found: ID does not exist" Oct 11 04:02:41 crc kubenswrapper[4967]: I1011 04:02:41.367868 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-console/console-f9d7485db-csl6r"] Oct 11 04:02:42 crc kubenswrapper[4967]: I1011 04:02:42.824470 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e20a512-8f34-4d21-be45-e749f7a6cc87" path="/var/lib/kubelet/pods/5e20a512-8f34-4d21-be45-e749f7a6cc87/volumes" Oct 11 04:02:43 crc kubenswrapper[4967]: I1011 04:02:43.336152 4967 generic.go:334] "Generic (PLEG): container finished" podID="bd662d9c-6ab0-4518-8fab-83115ccd3569" containerID="f127338a32c971b9fcedad564fce69857690938ae5f4ff01f028aeb0331f2fdc" exitCode=0 Oct 11 04:02:43 crc kubenswrapper[4967]: I1011 04:02:43.336565 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn" event={"ID":"bd662d9c-6ab0-4518-8fab-83115ccd3569","Type":"ContainerDied","Data":"f127338a32c971b9fcedad564fce69857690938ae5f4ff01f028aeb0331f2fdc"} Oct 11 04:02:44 crc kubenswrapper[4967]: I1011 04:02:44.088957 4967 patch_prober.go:28] interesting pod/machine-config-daemon-sgrv2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:02:44 crc kubenswrapper[4967]: I1011 04:02:44.089101 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:02:44 crc kubenswrapper[4967]: I1011 04:02:44.351898 4967 generic.go:334] "Generic (PLEG): container finished" podID="bd662d9c-6ab0-4518-8fab-83115ccd3569" containerID="2bc266893971d1a3a7289e9fca78a6b2f264b675bcfccb676e0925e601fdc44b" exitCode=0 Oct 11 04:02:44 crc kubenswrapper[4967]: I1011 04:02:44.351991 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn" event={"ID":"bd662d9c-6ab0-4518-8fab-83115ccd3569","Type":"ContainerDied","Data":"2bc266893971d1a3a7289e9fca78a6b2f264b675bcfccb676e0925e601fdc44b"} Oct 11 04:02:45 crc kubenswrapper[4967]: I1011 04:02:45.638913 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn" Oct 11 04:02:45 crc kubenswrapper[4967]: I1011 04:02:45.760125 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6fdg6\" (UniqueName: \"kubernetes.io/projected/bd662d9c-6ab0-4518-8fab-83115ccd3569-kube-api-access-6fdg6\") pod \"bd662d9c-6ab0-4518-8fab-83115ccd3569\" (UID: \"bd662d9c-6ab0-4518-8fab-83115ccd3569\") " Oct 11 04:02:45 crc kubenswrapper[4967]: I1011 04:02:45.760248 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd662d9c-6ab0-4518-8fab-83115ccd3569-bundle\") pod \"bd662d9c-6ab0-4518-8fab-83115ccd3569\" (UID: \"bd662d9c-6ab0-4518-8fab-83115ccd3569\") " Oct 11 04:02:45 crc kubenswrapper[4967]: I1011 04:02:45.760289 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd662d9c-6ab0-4518-8fab-83115ccd3569-util\") pod \"bd662d9c-6ab0-4518-8fab-83115ccd3569\" (UID: \"bd662d9c-6ab0-4518-8fab-83115ccd3569\") " Oct 11 04:02:45 crc kubenswrapper[4967]: I1011 04:02:45.761300 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd662d9c-6ab0-4518-8fab-83115ccd3569-bundle" (OuterVolumeSpecName: "bundle") pod "bd662d9c-6ab0-4518-8fab-83115ccd3569" (UID: "bd662d9c-6ab0-4518-8fab-83115ccd3569"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:02:45 crc kubenswrapper[4967]: I1011 04:02:45.766887 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd662d9c-6ab0-4518-8fab-83115ccd3569-kube-api-access-6fdg6" (OuterVolumeSpecName: "kube-api-access-6fdg6") pod "bd662d9c-6ab0-4518-8fab-83115ccd3569" (UID: "bd662d9c-6ab0-4518-8fab-83115ccd3569"). InnerVolumeSpecName "kube-api-access-6fdg6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:02:45 crc kubenswrapper[4967]: I1011 04:02:45.772944 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd662d9c-6ab0-4518-8fab-83115ccd3569-util" (OuterVolumeSpecName: "util") pod "bd662d9c-6ab0-4518-8fab-83115ccd3569" (UID: "bd662d9c-6ab0-4518-8fab-83115ccd3569"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:02:45 crc kubenswrapper[4967]: I1011 04:02:45.862217 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6fdg6\" (UniqueName: \"kubernetes.io/projected/bd662d9c-6ab0-4518-8fab-83115ccd3569-kube-api-access-6fdg6\") on node \"crc\" DevicePath \"\"" Oct 11 04:02:45 crc kubenswrapper[4967]: I1011 04:02:45.862267 4967 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/bd662d9c-6ab0-4518-8fab-83115ccd3569-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:02:45 crc kubenswrapper[4967]: I1011 04:02:45.862279 4967 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/bd662d9c-6ab0-4518-8fab-83115ccd3569-util\") on node \"crc\" DevicePath \"\"" Oct 11 04:02:46 crc kubenswrapper[4967]: I1011 04:02:46.366659 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn" event={"ID":"bd662d9c-6ab0-4518-8fab-83115ccd3569","Type":"ContainerDied","Data":"cee8826d98d53ea127cfad2a704c28c9a60d98d4d2fd234e11b0afa44ee1bda9"} Oct 11 04:02:46 crc kubenswrapper[4967]: I1011 04:02:46.366726 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cee8826d98d53ea127cfad2a704c28c9a60d98d4d2fd234e11b0afa44ee1bda9" Oct 11 04:02:46 crc kubenswrapper[4967]: I1011 04:02:46.366820 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.076979 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-controller-manager-5cfc444676-2s7hf"] Oct 11 04:02:54 crc kubenswrapper[4967]: E1011 04:02:54.077760 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd662d9c-6ab0-4518-8fab-83115ccd3569" containerName="extract" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.077774 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd662d9c-6ab0-4518-8fab-83115ccd3569" containerName="extract" Oct 11 04:02:54 crc kubenswrapper[4967]: E1011 04:02:54.077790 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd662d9c-6ab0-4518-8fab-83115ccd3569" containerName="util" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.077798 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd662d9c-6ab0-4518-8fab-83115ccd3569" containerName="util" Oct 11 04:02:54 crc kubenswrapper[4967]: E1011 04:02:54.077809 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd662d9c-6ab0-4518-8fab-83115ccd3569" containerName="pull" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.077815 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd662d9c-6ab0-4518-8fab-83115ccd3569" containerName="pull" Oct 11 04:02:54 crc kubenswrapper[4967]: E1011 04:02:54.077828 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e20a512-8f34-4d21-be45-e749f7a6cc87" containerName="console" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.077835 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e20a512-8f34-4d21-be45-e749f7a6cc87" containerName="console" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.077939 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd662d9c-6ab0-4518-8fab-83115ccd3569" containerName="extract" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.077951 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e20a512-8f34-4d21-be45-e749f7a6cc87" containerName="console" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.078429 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5cfc444676-2s7hf" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.096339 4967 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-cert" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.099492 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"openshift-service-ca.crt" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.099590 4967 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"manager-account-dockercfg-chg5m" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.099492 4967 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-controller-manager-service-cert" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.099511 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"kube-root-ca.crt" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.101463 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5cfc444676-2s7hf"] Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.172945 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hzmjf\" (UniqueName: \"kubernetes.io/projected/f3173c9b-89b4-420d-a852-f293d581da52-kube-api-access-hzmjf\") pod \"metallb-operator-controller-manager-5cfc444676-2s7hf\" (UID: \"f3173c9b-89b4-420d-a852-f293d581da52\") " pod="metallb-system/metallb-operator-controller-manager-5cfc444676-2s7hf" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.173083 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f3173c9b-89b4-420d-a852-f293d581da52-apiservice-cert\") pod \"metallb-operator-controller-manager-5cfc444676-2s7hf\" (UID: \"f3173c9b-89b4-420d-a852-f293d581da52\") " pod="metallb-system/metallb-operator-controller-manager-5cfc444676-2s7hf" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.173133 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f3173c9b-89b4-420d-a852-f293d581da52-webhook-cert\") pod \"metallb-operator-controller-manager-5cfc444676-2s7hf\" (UID: \"f3173c9b-89b4-420d-a852-f293d581da52\") " pod="metallb-system/metallb-operator-controller-manager-5cfc444676-2s7hf" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.274893 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f3173c9b-89b4-420d-a852-f293d581da52-apiservice-cert\") pod \"metallb-operator-controller-manager-5cfc444676-2s7hf\" (UID: \"f3173c9b-89b4-420d-a852-f293d581da52\") " pod="metallb-system/metallb-operator-controller-manager-5cfc444676-2s7hf" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.274947 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f3173c9b-89b4-420d-a852-f293d581da52-webhook-cert\") pod \"metallb-operator-controller-manager-5cfc444676-2s7hf\" (UID: \"f3173c9b-89b4-420d-a852-f293d581da52\") " pod="metallb-system/metallb-operator-controller-manager-5cfc444676-2s7hf" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.275005 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hzmjf\" (UniqueName: \"kubernetes.io/projected/f3173c9b-89b4-420d-a852-f293d581da52-kube-api-access-hzmjf\") pod \"metallb-operator-controller-manager-5cfc444676-2s7hf\" (UID: \"f3173c9b-89b4-420d-a852-f293d581da52\") " pod="metallb-system/metallb-operator-controller-manager-5cfc444676-2s7hf" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.281588 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/f3173c9b-89b4-420d-a852-f293d581da52-apiservice-cert\") pod \"metallb-operator-controller-manager-5cfc444676-2s7hf\" (UID: \"f3173c9b-89b4-420d-a852-f293d581da52\") " pod="metallb-system/metallb-operator-controller-manager-5cfc444676-2s7hf" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.291137 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/f3173c9b-89b4-420d-a852-f293d581da52-webhook-cert\") pod \"metallb-operator-controller-manager-5cfc444676-2s7hf\" (UID: \"f3173c9b-89b4-420d-a852-f293d581da52\") " pod="metallb-system/metallb-operator-controller-manager-5cfc444676-2s7hf" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.299106 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hzmjf\" (UniqueName: \"kubernetes.io/projected/f3173c9b-89b4-420d-a852-f293d581da52-kube-api-access-hzmjf\") pod \"metallb-operator-controller-manager-5cfc444676-2s7hf\" (UID: \"f3173c9b-89b4-420d-a852-f293d581da52\") " pod="metallb-system/metallb-operator-controller-manager-5cfc444676-2s7hf" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.369692 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/metallb-operator-webhook-server-56f68cd588-8fj48"] Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.370363 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-56f68cd588-8fj48" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.374870 4967 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.375053 4967 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-dockercfg-5tt68" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.377367 4967 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-operator-webhook-server-service-cert" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.400555 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-controller-manager-5cfc444676-2s7hf" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.471930 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-56f68cd588-8fj48"] Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.477045 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-n26zf\" (UniqueName: \"kubernetes.io/projected/c88642bc-96a7-4662-a035-c2c6043e322f-kube-api-access-n26zf\") pod \"metallb-operator-webhook-server-56f68cd588-8fj48\" (UID: \"c88642bc-96a7-4662-a035-c2c6043e322f\") " pod="metallb-system/metallb-operator-webhook-server-56f68cd588-8fj48" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.477151 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c88642bc-96a7-4662-a035-c2c6043e322f-apiservice-cert\") pod \"metallb-operator-webhook-server-56f68cd588-8fj48\" (UID: \"c88642bc-96a7-4662-a035-c2c6043e322f\") " pod="metallb-system/metallb-operator-webhook-server-56f68cd588-8fj48" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.477241 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c88642bc-96a7-4662-a035-c2c6043e322f-webhook-cert\") pod \"metallb-operator-webhook-server-56f68cd588-8fj48\" (UID: \"c88642bc-96a7-4662-a035-c2c6043e322f\") " pod="metallb-system/metallb-operator-webhook-server-56f68cd588-8fj48" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.578679 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-n26zf\" (UniqueName: \"kubernetes.io/projected/c88642bc-96a7-4662-a035-c2c6043e322f-kube-api-access-n26zf\") pod \"metallb-operator-webhook-server-56f68cd588-8fj48\" (UID: \"c88642bc-96a7-4662-a035-c2c6043e322f\") " pod="metallb-system/metallb-operator-webhook-server-56f68cd588-8fj48" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.579006 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c88642bc-96a7-4662-a035-c2c6043e322f-apiservice-cert\") pod \"metallb-operator-webhook-server-56f68cd588-8fj48\" (UID: \"c88642bc-96a7-4662-a035-c2c6043e322f\") " pod="metallb-system/metallb-operator-webhook-server-56f68cd588-8fj48" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.579048 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c88642bc-96a7-4662-a035-c2c6043e322f-webhook-cert\") pod \"metallb-operator-webhook-server-56f68cd588-8fj48\" (UID: \"c88642bc-96a7-4662-a035-c2c6043e322f\") " pod="metallb-system/metallb-operator-webhook-server-56f68cd588-8fj48" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.588698 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/c88642bc-96a7-4662-a035-c2c6043e322f-apiservice-cert\") pod \"metallb-operator-webhook-server-56f68cd588-8fj48\" (UID: \"c88642bc-96a7-4662-a035-c2c6043e322f\") " pod="metallb-system/metallb-operator-webhook-server-56f68cd588-8fj48" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.597641 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/c88642bc-96a7-4662-a035-c2c6043e322f-webhook-cert\") pod \"metallb-operator-webhook-server-56f68cd588-8fj48\" (UID: \"c88642bc-96a7-4662-a035-c2c6043e322f\") " pod="metallb-system/metallb-operator-webhook-server-56f68cd588-8fj48" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.604741 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-n26zf\" (UniqueName: \"kubernetes.io/projected/c88642bc-96a7-4662-a035-c2c6043e322f-kube-api-access-n26zf\") pod \"metallb-operator-webhook-server-56f68cd588-8fj48\" (UID: \"c88642bc-96a7-4662-a035-c2c6043e322f\") " pod="metallb-system/metallb-operator-webhook-server-56f68cd588-8fj48" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.691032 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/metallb-operator-webhook-server-56f68cd588-8fj48" Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.911930 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-controller-manager-5cfc444676-2s7hf"] Oct 11 04:02:54 crc kubenswrapper[4967]: W1011 04:02:54.914549 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc88642bc_96a7_4662_a035_c2c6043e322f.slice/crio-ad60ee5613e8d8a439b16ba457544ef662ecb299a6471ca8bb7dc8680207d5e4 WatchSource:0}: Error finding container ad60ee5613e8d8a439b16ba457544ef662ecb299a6471ca8bb7dc8680207d5e4: Status 404 returned error can't find the container with id ad60ee5613e8d8a439b16ba457544ef662ecb299a6471ca8bb7dc8680207d5e4 Oct 11 04:02:54 crc kubenswrapper[4967]: I1011 04:02:54.918966 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/metallb-operator-webhook-server-56f68cd588-8fj48"] Oct 11 04:02:55 crc kubenswrapper[4967]: I1011 04:02:55.416810 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-56f68cd588-8fj48" event={"ID":"c88642bc-96a7-4662-a035-c2c6043e322f","Type":"ContainerStarted","Data":"ad60ee5613e8d8a439b16ba457544ef662ecb299a6471ca8bb7dc8680207d5e4"} Oct 11 04:02:55 crc kubenswrapper[4967]: I1011 04:02:55.418208 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5cfc444676-2s7hf" event={"ID":"f3173c9b-89b4-420d-a852-f293d581da52","Type":"ContainerStarted","Data":"d4b33bdf57c630409610bc4581b266d0e560e41767bebe8849cbca06af85f503"} Oct 11 04:03:00 crc kubenswrapper[4967]: I1011 04:03:00.449789 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-controller-manager-5cfc444676-2s7hf" event={"ID":"f3173c9b-89b4-420d-a852-f293d581da52","Type":"ContainerStarted","Data":"bffc3ce9074218870cc487c454647bff3c70fd383c62b709b21a23fa1772c6f3"} Oct 11 04:03:00 crc kubenswrapper[4967]: I1011 04:03:00.450561 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-controller-manager-5cfc444676-2s7hf" Oct 11 04:03:00 crc kubenswrapper[4967]: I1011 04:03:00.452214 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/metallb-operator-webhook-server-56f68cd588-8fj48" event={"ID":"c88642bc-96a7-4662-a035-c2c6043e322f","Type":"ContainerStarted","Data":"5ed3921b334bbfaccd443f195bc4d29896dc37e65e3f6604e7efa4dee4c55eb0"} Oct 11 04:03:00 crc kubenswrapper[4967]: I1011 04:03:00.452439 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/metallb-operator-webhook-server-56f68cd588-8fj48" Oct 11 04:03:00 crc kubenswrapper[4967]: I1011 04:03:00.482704 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-controller-manager-5cfc444676-2s7hf" podStartSLOduration=2.027080592 podStartE2EDuration="6.482686418s" podCreationTimestamp="2025-10-11 04:02:54 +0000 UTC" firstStartedPulling="2025-10-11 04:02:54.912568643 +0000 UTC m=+702.875777576" lastFinishedPulling="2025-10-11 04:02:59.368174469 +0000 UTC m=+707.331383402" observedRunningTime="2025-10-11 04:03:00.478506654 +0000 UTC m=+708.441715597" watchObservedRunningTime="2025-10-11 04:03:00.482686418 +0000 UTC m=+708.445895361" Oct 11 04:03:00 crc kubenswrapper[4967]: I1011 04:03:00.515841 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/metallb-operator-webhook-server-56f68cd588-8fj48" podStartSLOduration=2.050144822 podStartE2EDuration="6.515821827s" podCreationTimestamp="2025-10-11 04:02:54 +0000 UTC" firstStartedPulling="2025-10-11 04:02:54.922543149 +0000 UTC m=+702.885752082" lastFinishedPulling="2025-10-11 04:02:59.388220134 +0000 UTC m=+707.351429087" observedRunningTime="2025-10-11 04:03:00.514410552 +0000 UTC m=+708.477619485" watchObservedRunningTime="2025-10-11 04:03:00.515821827 +0000 UTC m=+708.479030770" Oct 11 04:03:14 crc kubenswrapper[4967]: I1011 04:03:14.089420 4967 patch_prober.go:28] interesting pod/machine-config-daemon-sgrv2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:03:14 crc kubenswrapper[4967]: I1011 04:03:14.090842 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:03:14 crc kubenswrapper[4967]: I1011 04:03:14.695595 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-webhook-server-56f68cd588-8fj48" Oct 11 04:03:34 crc kubenswrapper[4967]: I1011 04:03:34.403928 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/metallb-operator-controller-manager-5cfc444676-2s7hf" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.148373 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-j9lg6"] Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.149561 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-j9lg6" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.152158 4967 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-daemon-dockercfg-wm45k" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.152470 4967 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-webhook-server-cert" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.165388 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-j9lg6"] Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.176188 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/frr-k8s-n67vr"] Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.178127 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-n67vr" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.180127 4967 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"frr-k8s-certs-secret" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.180956 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"frr-startup" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.235568 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e9d93993-50a0-468a-a3db-134db9acd6ba-reloader\") pod \"frr-k8s-n67vr\" (UID: \"e9d93993-50a0-468a-a3db-134db9acd6ba\") " pod="metallb-system/frr-k8s-n67vr" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.235763 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e9d93993-50a0-468a-a3db-134db9acd6ba-frr-startup\") pod \"frr-k8s-n67vr\" (UID: \"e9d93993-50a0-468a-a3db-134db9acd6ba\") " pod="metallb-system/frr-k8s-n67vr" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.235844 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e9d93993-50a0-468a-a3db-134db9acd6ba-frr-sockets\") pod \"frr-k8s-n67vr\" (UID: \"e9d93993-50a0-468a-a3db-134db9acd6ba\") " pod="metallb-system/frr-k8s-n67vr" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.235920 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e9d93993-50a0-468a-a3db-134db9acd6ba-frr-conf\") pod \"frr-k8s-n67vr\" (UID: \"e9d93993-50a0-468a-a3db-134db9acd6ba\") " pod="metallb-system/frr-k8s-n67vr" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.236018 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e9d93993-50a0-468a-a3db-134db9acd6ba-metrics\") pod \"frr-k8s-n67vr\" (UID: \"e9d93993-50a0-468a-a3db-134db9acd6ba\") " pod="metallb-system/frr-k8s-n67vr" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.244840 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/speaker-np2lm"] Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.245854 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-np2lm" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.251594 4967 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-memberlist" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.251863 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"metallb-system"/"metallb-excludel2" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.252791 4967 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-certs-secret" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.252840 4967 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"speaker-dockercfg-pqbb5" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.266008 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["metallb-system/controller-68d546b9d8-rv747"] Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.267018 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-rv747" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.269337 4967 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"controller-certs-secret" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.281481 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-rv747"] Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.337002 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e9d93993-50a0-468a-a3db-134db9acd6ba-reloader\") pod \"frr-k8s-n67vr\" (UID: \"e9d93993-50a0-468a-a3db-134db9acd6ba\") " pod="metallb-system/frr-k8s-n67vr" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.337047 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e9d93993-50a0-468a-a3db-134db9acd6ba-frr-startup\") pod \"frr-k8s-n67vr\" (UID: \"e9d93993-50a0-468a-a3db-134db9acd6ba\") " pod="metallb-system/frr-k8s-n67vr" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.337094 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e9d93993-50a0-468a-a3db-134db9acd6ba-frr-sockets\") pod \"frr-k8s-n67vr\" (UID: \"e9d93993-50a0-468a-a3db-134db9acd6ba\") " pod="metallb-system/frr-k8s-n67vr" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.337118 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/50110bd6-5363-43c8-b3a8-b4b3762c0c43-cert\") pod \"frr-k8s-webhook-server-64bf5d555-j9lg6\" (UID: \"50110bd6-5363-43c8-b3a8-b4b3762c0c43\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-j9lg6" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.337141 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e9d93993-50a0-468a-a3db-134db9acd6ba-frr-conf\") pod \"frr-k8s-n67vr\" (UID: \"e9d93993-50a0-468a-a3db-134db9acd6ba\") " pod="metallb-system/frr-k8s-n67vr" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.337187 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e9d93993-50a0-468a-a3db-134db9acd6ba-metrics\") pod \"frr-k8s-n67vr\" (UID: \"e9d93993-50a0-468a-a3db-134db9acd6ba\") " pod="metallb-system/frr-k8s-n67vr" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.337214 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9d93993-50a0-468a-a3db-134db9acd6ba-metrics-certs\") pod \"frr-k8s-n67vr\" (UID: \"e9d93993-50a0-468a-a3db-134db9acd6ba\") " pod="metallb-system/frr-k8s-n67vr" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.337242 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jn4z\" (UniqueName: \"kubernetes.io/projected/50110bd6-5363-43c8-b3a8-b4b3762c0c43-kube-api-access-8jn4z\") pod \"frr-k8s-webhook-server-64bf5d555-j9lg6\" (UID: \"50110bd6-5363-43c8-b3a8-b4b3762c0c43\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-j9lg6" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.337273 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qgktg\" (UniqueName: \"kubernetes.io/projected/e9d93993-50a0-468a-a3db-134db9acd6ba-kube-api-access-qgktg\") pod \"frr-k8s-n67vr\" (UID: \"e9d93993-50a0-468a-a3db-134db9acd6ba\") " pod="metallb-system/frr-k8s-n67vr" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.337713 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"reloader\" (UniqueName: \"kubernetes.io/empty-dir/e9d93993-50a0-468a-a3db-134db9acd6ba-reloader\") pod \"frr-k8s-n67vr\" (UID: \"e9d93993-50a0-468a-a3db-134db9acd6ba\") " pod="metallb-system/frr-k8s-n67vr" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.338329 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-conf\" (UniqueName: \"kubernetes.io/empty-dir/e9d93993-50a0-468a-a3db-134db9acd6ba-frr-conf\") pod \"frr-k8s-n67vr\" (UID: \"e9d93993-50a0-468a-a3db-134db9acd6ba\") " pod="metallb-system/frr-k8s-n67vr" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.338541 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-sockets\" (UniqueName: \"kubernetes.io/empty-dir/e9d93993-50a0-468a-a3db-134db9acd6ba-frr-sockets\") pod \"frr-k8s-n67vr\" (UID: \"e9d93993-50a0-468a-a3db-134db9acd6ba\") " pod="metallb-system/frr-k8s-n67vr" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.338694 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics\" (UniqueName: \"kubernetes.io/empty-dir/e9d93993-50a0-468a-a3db-134db9acd6ba-metrics\") pod \"frr-k8s-n67vr\" (UID: \"e9d93993-50a0-468a-a3db-134db9acd6ba\") " pod="metallb-system/frr-k8s-n67vr" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.339169 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"frr-startup\" (UniqueName: \"kubernetes.io/configmap/e9d93993-50a0-468a-a3db-134db9acd6ba-frr-startup\") pod \"frr-k8s-n67vr\" (UID: \"e9d93993-50a0-468a-a3db-134db9acd6ba\") " pod="metallb-system/frr-k8s-n67vr" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.437832 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9d93993-50a0-468a-a3db-134db9acd6ba-metrics-certs\") pod \"frr-k8s-n67vr\" (UID: \"e9d93993-50a0-468a-a3db-134db9acd6ba\") " pod="metallb-system/frr-k8s-n67vr" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.437886 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j8ggk\" (UniqueName: \"kubernetes.io/projected/816b70d7-ff40-4cee-b21d-01008f0d4868-kube-api-access-j8ggk\") pod \"speaker-np2lm\" (UID: \"816b70d7-ff40-4cee-b21d-01008f0d4868\") " pod="metallb-system/speaker-np2lm" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.437921 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jn4z\" (UniqueName: \"kubernetes.io/projected/50110bd6-5363-43c8-b3a8-b4b3762c0c43-kube-api-access-8jn4z\") pod \"frr-k8s-webhook-server-64bf5d555-j9lg6\" (UID: \"50110bd6-5363-43c8-b3a8-b4b3762c0c43\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-j9lg6" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.437944 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/816b70d7-ff40-4cee-b21d-01008f0d4868-metallb-excludel2\") pod \"speaker-np2lm\" (UID: \"816b70d7-ff40-4cee-b21d-01008f0d4868\") " pod="metallb-system/speaker-np2lm" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.437965 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qgktg\" (UniqueName: \"kubernetes.io/projected/e9d93993-50a0-468a-a3db-134db9acd6ba-kube-api-access-qgktg\") pod \"frr-k8s-n67vr\" (UID: \"e9d93993-50a0-468a-a3db-134db9acd6ba\") " pod="metallb-system/frr-k8s-n67vr" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.438000 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7fb0bb1d-4f24-4ca2-8148-ac03ee0083c5-cert\") pod \"controller-68d546b9d8-rv747\" (UID: \"7fb0bb1d-4f24-4ca2-8148-ac03ee0083c5\") " pod="metallb-system/controller-68d546b9d8-rv747" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.438020 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jbkd5\" (UniqueName: \"kubernetes.io/projected/7fb0bb1d-4f24-4ca2-8148-ac03ee0083c5-kube-api-access-jbkd5\") pod \"controller-68d546b9d8-rv747\" (UID: \"7fb0bb1d-4f24-4ca2-8148-ac03ee0083c5\") " pod="metallb-system/controller-68d546b9d8-rv747" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.438053 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/50110bd6-5363-43c8-b3a8-b4b3762c0c43-cert\") pod \"frr-k8s-webhook-server-64bf5d555-j9lg6\" (UID: \"50110bd6-5363-43c8-b3a8-b4b3762c0c43\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-j9lg6" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.438088 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7fb0bb1d-4f24-4ca2-8148-ac03ee0083c5-metrics-certs\") pod \"controller-68d546b9d8-rv747\" (UID: \"7fb0bb1d-4f24-4ca2-8148-ac03ee0083c5\") " pod="metallb-system/controller-68d546b9d8-rv747" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.438115 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/816b70d7-ff40-4cee-b21d-01008f0d4868-memberlist\") pod \"speaker-np2lm\" (UID: \"816b70d7-ff40-4cee-b21d-01008f0d4868\") " pod="metallb-system/speaker-np2lm" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.438133 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/816b70d7-ff40-4cee-b21d-01008f0d4868-metrics-certs\") pod \"speaker-np2lm\" (UID: \"816b70d7-ff40-4cee-b21d-01008f0d4868\") " pod="metallb-system/speaker-np2lm" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.444785 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/50110bd6-5363-43c8-b3a8-b4b3762c0c43-cert\") pod \"frr-k8s-webhook-server-64bf5d555-j9lg6\" (UID: \"50110bd6-5363-43c8-b3a8-b4b3762c0c43\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-j9lg6" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.448664 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e9d93993-50a0-468a-a3db-134db9acd6ba-metrics-certs\") pod \"frr-k8s-n67vr\" (UID: \"e9d93993-50a0-468a-a3db-134db9acd6ba\") " pod="metallb-system/frr-k8s-n67vr" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.454898 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qgktg\" (UniqueName: \"kubernetes.io/projected/e9d93993-50a0-468a-a3db-134db9acd6ba-kube-api-access-qgktg\") pod \"frr-k8s-n67vr\" (UID: \"e9d93993-50a0-468a-a3db-134db9acd6ba\") " pod="metallb-system/frr-k8s-n67vr" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.456395 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jn4z\" (UniqueName: \"kubernetes.io/projected/50110bd6-5363-43c8-b3a8-b4b3762c0c43-kube-api-access-8jn4z\") pod \"frr-k8s-webhook-server-64bf5d555-j9lg6\" (UID: \"50110bd6-5363-43c8-b3a8-b4b3762c0c43\") " pod="metallb-system/frr-k8s-webhook-server-64bf5d555-j9lg6" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.479198 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-j9lg6" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.499227 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/frr-k8s-n67vr" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.543721 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7fb0bb1d-4f24-4ca2-8148-ac03ee0083c5-metrics-certs\") pod \"controller-68d546b9d8-rv747\" (UID: \"7fb0bb1d-4f24-4ca2-8148-ac03ee0083c5\") " pod="metallb-system/controller-68d546b9d8-rv747" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.543792 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/816b70d7-ff40-4cee-b21d-01008f0d4868-memberlist\") pod \"speaker-np2lm\" (UID: \"816b70d7-ff40-4cee-b21d-01008f0d4868\") " pod="metallb-system/speaker-np2lm" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.543810 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/816b70d7-ff40-4cee-b21d-01008f0d4868-metrics-certs\") pod \"speaker-np2lm\" (UID: \"816b70d7-ff40-4cee-b21d-01008f0d4868\") " pod="metallb-system/speaker-np2lm" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.543840 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j8ggk\" (UniqueName: \"kubernetes.io/projected/816b70d7-ff40-4cee-b21d-01008f0d4868-kube-api-access-j8ggk\") pod \"speaker-np2lm\" (UID: \"816b70d7-ff40-4cee-b21d-01008f0d4868\") " pod="metallb-system/speaker-np2lm" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.543863 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/816b70d7-ff40-4cee-b21d-01008f0d4868-metallb-excludel2\") pod \"speaker-np2lm\" (UID: \"816b70d7-ff40-4cee-b21d-01008f0d4868\") " pod="metallb-system/speaker-np2lm" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.543891 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7fb0bb1d-4f24-4ca2-8148-ac03ee0083c5-cert\") pod \"controller-68d546b9d8-rv747\" (UID: \"7fb0bb1d-4f24-4ca2-8148-ac03ee0083c5\") " pod="metallb-system/controller-68d546b9d8-rv747" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.543910 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jbkd5\" (UniqueName: \"kubernetes.io/projected/7fb0bb1d-4f24-4ca2-8148-ac03ee0083c5-kube-api-access-jbkd5\") pod \"controller-68d546b9d8-rv747\" (UID: \"7fb0bb1d-4f24-4ca2-8148-ac03ee0083c5\") " pod="metallb-system/controller-68d546b9d8-rv747" Oct 11 04:03:35 crc kubenswrapper[4967]: E1011 04:03:35.544435 4967 secret.go:188] Couldn't get secret metallb-system/metallb-memberlist: secret "metallb-memberlist" not found Oct 11 04:03:35 crc kubenswrapper[4967]: E1011 04:03:35.544508 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/816b70d7-ff40-4cee-b21d-01008f0d4868-memberlist podName:816b70d7-ff40-4cee-b21d-01008f0d4868 nodeName:}" failed. No retries permitted until 2025-10-11 04:03:36.044487537 +0000 UTC m=+744.007696470 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "memberlist" (UniqueName: "kubernetes.io/secret/816b70d7-ff40-4cee-b21d-01008f0d4868-memberlist") pod "speaker-np2lm" (UID: "816b70d7-ff40-4cee-b21d-01008f0d4868") : secret "metallb-memberlist" not found Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.545013 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metallb-excludel2\" (UniqueName: \"kubernetes.io/configmap/816b70d7-ff40-4cee-b21d-01008f0d4868-metallb-excludel2\") pod \"speaker-np2lm\" (UID: \"816b70d7-ff40-4cee-b21d-01008f0d4868\") " pod="metallb-system/speaker-np2lm" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.550428 4967 reflector.go:368] Caches populated for *v1.Secret from object-"metallb-system"/"metallb-webhook-cert" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.550889 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/7fb0bb1d-4f24-4ca2-8148-ac03ee0083c5-metrics-certs\") pod \"controller-68d546b9d8-rv747\" (UID: \"7fb0bb1d-4f24-4ca2-8148-ac03ee0083c5\") " pod="metallb-system/controller-68d546b9d8-rv747" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.553555 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/816b70d7-ff40-4cee-b21d-01008f0d4868-metrics-certs\") pod \"speaker-np2lm\" (UID: \"816b70d7-ff40-4cee-b21d-01008f0d4868\") " pod="metallb-system/speaker-np2lm" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.559791 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/7fb0bb1d-4f24-4ca2-8148-ac03ee0083c5-cert\") pod \"controller-68d546b9d8-rv747\" (UID: \"7fb0bb1d-4f24-4ca2-8148-ac03ee0083c5\") " pod="metallb-system/controller-68d546b9d8-rv747" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.562589 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jbkd5\" (UniqueName: \"kubernetes.io/projected/7fb0bb1d-4f24-4ca2-8148-ac03ee0083c5-kube-api-access-jbkd5\") pod \"controller-68d546b9d8-rv747\" (UID: \"7fb0bb1d-4f24-4ca2-8148-ac03ee0083c5\") " pod="metallb-system/controller-68d546b9d8-rv747" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.563501 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j8ggk\" (UniqueName: \"kubernetes.io/projected/816b70d7-ff40-4cee-b21d-01008f0d4868-kube-api-access-j8ggk\") pod \"speaker-np2lm\" (UID: \"816b70d7-ff40-4cee-b21d-01008f0d4868\") " pod="metallb-system/speaker-np2lm" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.584993 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/controller-68d546b9d8-rv747" Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.713130 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-n67vr" event={"ID":"e9d93993-50a0-468a-a3db-134db9acd6ba","Type":"ContainerStarted","Data":"479fb69ea07a586f3697167fbbd21e14226218eecc6691dd4004344fecb723e9"} Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.806517 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/controller-68d546b9d8-rv747"] Oct 11 04:03:35 crc kubenswrapper[4967]: W1011 04:03:35.814034 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7fb0bb1d_4f24_4ca2_8148_ac03ee0083c5.slice/crio-4af3fbb548dbcf67de7a0ba9dd707ebd1408fe03c1d7ffa0c18c3f51acd62fc0 WatchSource:0}: Error finding container 4af3fbb548dbcf67de7a0ba9dd707ebd1408fe03c1d7ffa0c18c3f51acd62fc0: Status 404 returned error can't find the container with id 4af3fbb548dbcf67de7a0ba9dd707ebd1408fe03c1d7ffa0c18c3f51acd62fc0 Oct 11 04:03:35 crc kubenswrapper[4967]: I1011 04:03:35.961101 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["metallb-system/frr-k8s-webhook-server-64bf5d555-j9lg6"] Oct 11 04:03:36 crc kubenswrapper[4967]: I1011 04:03:36.052206 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/816b70d7-ff40-4cee-b21d-01008f0d4868-memberlist\") pod \"speaker-np2lm\" (UID: \"816b70d7-ff40-4cee-b21d-01008f0d4868\") " pod="metallb-system/speaker-np2lm" Oct 11 04:03:36 crc kubenswrapper[4967]: I1011 04:03:36.058038 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memberlist\" (UniqueName: \"kubernetes.io/secret/816b70d7-ff40-4cee-b21d-01008f0d4868-memberlist\") pod \"speaker-np2lm\" (UID: \"816b70d7-ff40-4cee-b21d-01008f0d4868\") " pod="metallb-system/speaker-np2lm" Oct 11 04:03:36 crc kubenswrapper[4967]: I1011 04:03:36.159324 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="metallb-system/speaker-np2lm" Oct 11 04:03:36 crc kubenswrapper[4967]: W1011 04:03:36.185656 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod816b70d7_ff40_4cee_b21d_01008f0d4868.slice/crio-002c1057e76a303d4da01989504a01ac710001c04ca21948b0a40c756dac39f6 WatchSource:0}: Error finding container 002c1057e76a303d4da01989504a01ac710001c04ca21948b0a40c756dac39f6: Status 404 returned error can't find the container with id 002c1057e76a303d4da01989504a01ac710001c04ca21948b0a40c756dac39f6 Oct 11 04:03:36 crc kubenswrapper[4967]: I1011 04:03:36.722616 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-j9lg6" event={"ID":"50110bd6-5363-43c8-b3a8-b4b3762c0c43","Type":"ContainerStarted","Data":"99bf77022b9eff075d134ceedbf26b32abaa78ed400b711f1d30b66436823ede"} Oct 11 04:03:36 crc kubenswrapper[4967]: I1011 04:03:36.728056 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-np2lm" event={"ID":"816b70d7-ff40-4cee-b21d-01008f0d4868","Type":"ContainerStarted","Data":"42becceeb97e0738d6e2d0322737dc0e1cae829b315fcfe196f76196338f0bbd"} Oct 11 04:03:36 crc kubenswrapper[4967]: I1011 04:03:36.728118 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-np2lm" event={"ID":"816b70d7-ff40-4cee-b21d-01008f0d4868","Type":"ContainerStarted","Data":"2572de50d0131b70b3b926623ec18225bfee1dd2a1a54546853e106312f0bdb6"} Oct 11 04:03:36 crc kubenswrapper[4967]: I1011 04:03:36.728130 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/speaker-np2lm" event={"ID":"816b70d7-ff40-4cee-b21d-01008f0d4868","Type":"ContainerStarted","Data":"002c1057e76a303d4da01989504a01ac710001c04ca21948b0a40c756dac39f6"} Oct 11 04:03:36 crc kubenswrapper[4967]: I1011 04:03:36.728305 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/speaker-np2lm" Oct 11 04:03:36 crc kubenswrapper[4967]: I1011 04:03:36.733546 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-rv747" event={"ID":"7fb0bb1d-4f24-4ca2-8148-ac03ee0083c5","Type":"ContainerStarted","Data":"8875123ccccda390ed5a2491e4db1ea25bacc8b7fb871b01279d5eb9010e3c11"} Oct 11 04:03:36 crc kubenswrapper[4967]: I1011 04:03:36.733589 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-rv747" event={"ID":"7fb0bb1d-4f24-4ca2-8148-ac03ee0083c5","Type":"ContainerStarted","Data":"1d4c6763cdb388f27d8230db032ec45febb831cacf3512acff4c45ebb0db8f33"} Oct 11 04:03:36 crc kubenswrapper[4967]: I1011 04:03:36.733599 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/controller-68d546b9d8-rv747" event={"ID":"7fb0bb1d-4f24-4ca2-8148-ac03ee0083c5","Type":"ContainerStarted","Data":"4af3fbb548dbcf67de7a0ba9dd707ebd1408fe03c1d7ffa0c18c3f51acd62fc0"} Oct 11 04:03:36 crc kubenswrapper[4967]: I1011 04:03:36.733711 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/controller-68d546b9d8-rv747" Oct 11 04:03:36 crc kubenswrapper[4967]: I1011 04:03:36.751889 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/speaker-np2lm" podStartSLOduration=1.751872306 podStartE2EDuration="1.751872306s" podCreationTimestamp="2025-10-11 04:03:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:03:36.750506934 +0000 UTC m=+744.713715867" watchObservedRunningTime="2025-10-11 04:03:36.751872306 +0000 UTC m=+744.715081239" Oct 11 04:03:36 crc kubenswrapper[4967]: I1011 04:03:36.784133 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/controller-68d546b9d8-rv747" podStartSLOduration=1.784114285 podStartE2EDuration="1.784114285s" podCreationTimestamp="2025-10-11 04:03:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:03:36.777535162 +0000 UTC m=+744.740744115" watchObservedRunningTime="2025-10-11 04:03:36.784114285 +0000 UTC m=+744.747323218" Oct 11 04:03:39 crc kubenswrapper[4967]: I1011 04:03:39.609052 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-92sz4"] Oct 11 04:03:39 crc kubenswrapper[4967]: I1011 04:03:39.609565 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-92sz4" podUID="1c77fe9a-d2e1-465b-8600-9f7f2d1fa206" containerName="controller-manager" containerID="cri-o://f8cb425ed8afecfff544a72cae16fd56f27a0a17fdf2156e2e4dae6ab1ba7fc4" gracePeriod=30 Oct 11 04:03:39 crc kubenswrapper[4967]: I1011 04:03:39.715950 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wtn79"] Oct 11 04:03:39 crc kubenswrapper[4967]: I1011 04:03:39.716182 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wtn79" podUID="5449c948-487e-426d-b151-31e8079576f3" containerName="route-controller-manager" containerID="cri-o://71fb1ca53ba385ae5087477e3e0fee48dbe734d830c5e65433013adbbd6f396c" gracePeriod=30 Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.123160 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-92sz4" Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.128512 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wtn79" Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.212539 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c77fe9a-d2e1-465b-8600-9f7f2d1fa206-serving-cert\") pod \"1c77fe9a-d2e1-465b-8600-9f7f2d1fa206\" (UID: \"1c77fe9a-d2e1-465b-8600-9f7f2d1fa206\") " Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.212588 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1c77fe9a-d2e1-465b-8600-9f7f2d1fa206-proxy-ca-bundles\") pod \"1c77fe9a-d2e1-465b-8600-9f7f2d1fa206\" (UID: \"1c77fe9a-d2e1-465b-8600-9f7f2d1fa206\") " Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.212631 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5449c948-487e-426d-b151-31e8079576f3-config\") pod \"5449c948-487e-426d-b151-31e8079576f3\" (UID: \"5449c948-487e-426d-b151-31e8079576f3\") " Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.212659 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5449c948-487e-426d-b151-31e8079576f3-serving-cert\") pod \"5449c948-487e-426d-b151-31e8079576f3\" (UID: \"5449c948-487e-426d-b151-31e8079576f3\") " Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.212689 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qvzlg\" (UniqueName: \"kubernetes.io/projected/5449c948-487e-426d-b151-31e8079576f3-kube-api-access-qvzlg\") pod \"5449c948-487e-426d-b151-31e8079576f3\" (UID: \"5449c948-487e-426d-b151-31e8079576f3\") " Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.212715 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1c77fe9a-d2e1-465b-8600-9f7f2d1fa206-client-ca\") pod \"1c77fe9a-d2e1-465b-8600-9f7f2d1fa206\" (UID: \"1c77fe9a-d2e1-465b-8600-9f7f2d1fa206\") " Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.212734 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c77fe9a-d2e1-465b-8600-9f7f2d1fa206-config\") pod \"1c77fe9a-d2e1-465b-8600-9f7f2d1fa206\" (UID: \"1c77fe9a-d2e1-465b-8600-9f7f2d1fa206\") " Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.212798 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5449c948-487e-426d-b151-31e8079576f3-client-ca\") pod \"5449c948-487e-426d-b151-31e8079576f3\" (UID: \"5449c948-487e-426d-b151-31e8079576f3\") " Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.212835 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-j8vnp\" (UniqueName: \"kubernetes.io/projected/1c77fe9a-d2e1-465b-8600-9f7f2d1fa206-kube-api-access-j8vnp\") pod \"1c77fe9a-d2e1-465b-8600-9f7f2d1fa206\" (UID: \"1c77fe9a-d2e1-465b-8600-9f7f2d1fa206\") " Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.213464 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c77fe9a-d2e1-465b-8600-9f7f2d1fa206-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "1c77fe9a-d2e1-465b-8600-9f7f2d1fa206" (UID: "1c77fe9a-d2e1-465b-8600-9f7f2d1fa206"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.216437 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c77fe9a-d2e1-465b-8600-9f7f2d1fa206-config" (OuterVolumeSpecName: "config") pod "1c77fe9a-d2e1-465b-8600-9f7f2d1fa206" (UID: "1c77fe9a-d2e1-465b-8600-9f7f2d1fa206"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.218400 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5449c948-487e-426d-b151-31e8079576f3-kube-api-access-qvzlg" (OuterVolumeSpecName: "kube-api-access-qvzlg") pod "5449c948-487e-426d-b151-31e8079576f3" (UID: "5449c948-487e-426d-b151-31e8079576f3"). InnerVolumeSpecName "kube-api-access-qvzlg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.219150 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5449c948-487e-426d-b151-31e8079576f3-config" (OuterVolumeSpecName: "config") pod "5449c948-487e-426d-b151-31e8079576f3" (UID: "5449c948-487e-426d-b151-31e8079576f3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.219216 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5449c948-487e-426d-b151-31e8079576f3-client-ca" (OuterVolumeSpecName: "client-ca") pod "5449c948-487e-426d-b151-31e8079576f3" (UID: "5449c948-487e-426d-b151-31e8079576f3"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.219424 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1c77fe9a-d2e1-465b-8600-9f7f2d1fa206-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1c77fe9a-d2e1-465b-8600-9f7f2d1fa206" (UID: "1c77fe9a-d2e1-465b-8600-9f7f2d1fa206"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.219554 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1c77fe9a-d2e1-465b-8600-9f7f2d1fa206-kube-api-access-j8vnp" (OuterVolumeSpecName: "kube-api-access-j8vnp") pod "1c77fe9a-d2e1-465b-8600-9f7f2d1fa206" (UID: "1c77fe9a-d2e1-465b-8600-9f7f2d1fa206"). InnerVolumeSpecName "kube-api-access-j8vnp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.221267 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5449c948-487e-426d-b151-31e8079576f3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5449c948-487e-426d-b151-31e8079576f3" (UID: "5449c948-487e-426d-b151-31e8079576f3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.224273 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1c77fe9a-d2e1-465b-8600-9f7f2d1fa206-client-ca" (OuterVolumeSpecName: "client-ca") pod "1c77fe9a-d2e1-465b-8600-9f7f2d1fa206" (UID: "1c77fe9a-d2e1-465b-8600-9f7f2d1fa206"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.314539 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5449c948-487e-426d-b151-31e8079576f3-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.314850 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qvzlg\" (UniqueName: \"kubernetes.io/projected/5449c948-487e-426d-b151-31e8079576f3-kube-api-access-qvzlg\") on node \"crc\" DevicePath \"\"" Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.314862 4967 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/1c77fe9a-d2e1-465b-8600-9f7f2d1fa206-client-ca\") on node \"crc\" DevicePath \"\"" Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.314871 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1c77fe9a-d2e1-465b-8600-9f7f2d1fa206-config\") on node \"crc\" DevicePath \"\"" Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.314879 4967 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5449c948-487e-426d-b151-31e8079576f3-client-ca\") on node \"crc\" DevicePath \"\"" Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.314887 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-j8vnp\" (UniqueName: \"kubernetes.io/projected/1c77fe9a-d2e1-465b-8600-9f7f2d1fa206-kube-api-access-j8vnp\") on node \"crc\" DevicePath \"\"" Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.314895 4967 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1c77fe9a-d2e1-465b-8600-9f7f2d1fa206-serving-cert\") on node \"crc\" DevicePath \"\"" Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.314903 4967 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/1c77fe9a-d2e1-465b-8600-9f7f2d1fa206-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.314912 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5449c948-487e-426d-b151-31e8079576f3-config\") on node \"crc\" DevicePath \"\"" Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.759916 4967 generic.go:334] "Generic (PLEG): container finished" podID="5449c948-487e-426d-b151-31e8079576f3" containerID="71fb1ca53ba385ae5087477e3e0fee48dbe734d830c5e65433013adbbd6f396c" exitCode=0 Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.759980 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wtn79" event={"ID":"5449c948-487e-426d-b151-31e8079576f3","Type":"ContainerDied","Data":"71fb1ca53ba385ae5087477e3e0fee48dbe734d830c5e65433013adbbd6f396c"} Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.760005 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wtn79" event={"ID":"5449c948-487e-426d-b151-31e8079576f3","Type":"ContainerDied","Data":"f540a51e3cfc72196043bc956e60b32300dabd6c5ae10b9d0a1dd1a07b86aff6"} Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.760022 4967 scope.go:117] "RemoveContainer" containerID="71fb1ca53ba385ae5087477e3e0fee48dbe734d830c5e65433013adbbd6f396c" Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.760020 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-wtn79" Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.764375 4967 generic.go:334] "Generic (PLEG): container finished" podID="1c77fe9a-d2e1-465b-8600-9f7f2d1fa206" containerID="f8cb425ed8afecfff544a72cae16fd56f27a0a17fdf2156e2e4dae6ab1ba7fc4" exitCode=0 Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.764409 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-92sz4" event={"ID":"1c77fe9a-d2e1-465b-8600-9f7f2d1fa206","Type":"ContainerDied","Data":"f8cb425ed8afecfff544a72cae16fd56f27a0a17fdf2156e2e4dae6ab1ba7fc4"} Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.764434 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-92sz4" event={"ID":"1c77fe9a-d2e1-465b-8600-9f7f2d1fa206","Type":"ContainerDied","Data":"8675633416cc7576fe45dc4505545a7520fc1778845f60d19e362df7a3e1480a"} Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.764498 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-92sz4" Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.796372 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wtn79"] Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.801022 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-wtn79"] Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.805507 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-92sz4"] Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.807257 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-92sz4"] Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.828504 4967 scope.go:117] "RemoveContainer" containerID="71fb1ca53ba385ae5087477e3e0fee48dbe734d830c5e65433013adbbd6f396c" Oct 11 04:03:40 crc kubenswrapper[4967]: E1011 04:03:40.828962 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"71fb1ca53ba385ae5087477e3e0fee48dbe734d830c5e65433013adbbd6f396c\": container with ID starting with 71fb1ca53ba385ae5087477e3e0fee48dbe734d830c5e65433013adbbd6f396c not found: ID does not exist" containerID="71fb1ca53ba385ae5087477e3e0fee48dbe734d830c5e65433013adbbd6f396c" Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.829009 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"71fb1ca53ba385ae5087477e3e0fee48dbe734d830c5e65433013adbbd6f396c"} err="failed to get container status \"71fb1ca53ba385ae5087477e3e0fee48dbe734d830c5e65433013adbbd6f396c\": rpc error: code = NotFound desc = could not find container \"71fb1ca53ba385ae5087477e3e0fee48dbe734d830c5e65433013adbbd6f396c\": container with ID starting with 71fb1ca53ba385ae5087477e3e0fee48dbe734d830c5e65433013adbbd6f396c not found: ID does not exist" Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.829035 4967 scope.go:117] "RemoveContainer" containerID="f8cb425ed8afecfff544a72cae16fd56f27a0a17fdf2156e2e4dae6ab1ba7fc4" Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.832207 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1c77fe9a-d2e1-465b-8600-9f7f2d1fa206" path="/var/lib/kubelet/pods/1c77fe9a-d2e1-465b-8600-9f7f2d1fa206/volumes" Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.832821 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5449c948-487e-426d-b151-31e8079576f3" path="/var/lib/kubelet/pods/5449c948-487e-426d-b151-31e8079576f3/volumes" Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.845443 4967 scope.go:117] "RemoveContainer" containerID="f8cb425ed8afecfff544a72cae16fd56f27a0a17fdf2156e2e4dae6ab1ba7fc4" Oct 11 04:03:40 crc kubenswrapper[4967]: E1011 04:03:40.845924 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f8cb425ed8afecfff544a72cae16fd56f27a0a17fdf2156e2e4dae6ab1ba7fc4\": container with ID starting with f8cb425ed8afecfff544a72cae16fd56f27a0a17fdf2156e2e4dae6ab1ba7fc4 not found: ID does not exist" containerID="f8cb425ed8afecfff544a72cae16fd56f27a0a17fdf2156e2e4dae6ab1ba7fc4" Oct 11 04:03:40 crc kubenswrapper[4967]: I1011 04:03:40.846019 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f8cb425ed8afecfff544a72cae16fd56f27a0a17fdf2156e2e4dae6ab1ba7fc4"} err="failed to get container status \"f8cb425ed8afecfff544a72cae16fd56f27a0a17fdf2156e2e4dae6ab1ba7fc4\": rpc error: code = NotFound desc = could not find container \"f8cb425ed8afecfff544a72cae16fd56f27a0a17fdf2156e2e4dae6ab1ba7fc4\": container with ID starting with f8cb425ed8afecfff544a72cae16fd56f27a0a17fdf2156e2e4dae6ab1ba7fc4 not found: ID does not exist" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.149641 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-77bc48855d-v7rlc"] Oct 11 04:03:41 crc kubenswrapper[4967]: E1011 04:03:41.149897 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1c77fe9a-d2e1-465b-8600-9f7f2d1fa206" containerName="controller-manager" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.149912 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="1c77fe9a-d2e1-465b-8600-9f7f2d1fa206" containerName="controller-manager" Oct 11 04:03:41 crc kubenswrapper[4967]: E1011 04:03:41.149927 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5449c948-487e-426d-b151-31e8079576f3" containerName="route-controller-manager" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.149933 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="5449c948-487e-426d-b151-31e8079576f3" containerName="route-controller-manager" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.150062 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="5449c948-487e-426d-b151-31e8079576f3" containerName="route-controller-manager" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.150103 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="1c77fe9a-d2e1-465b-8600-9f7f2d1fa206" containerName="controller-manager" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.150556 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-77bc48855d-v7rlc" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.152662 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Oct 11 04:03:41 crc kubenswrapper[4967]: W1011 04:03:41.152941 4967 reflector.go:561] object-"openshift-controller-manager"/"serving-cert": failed to list *v1.Secret: secrets "serving-cert" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openshift-controller-manager": no relationship found between node 'crc' and this object Oct 11 04:03:41 crc kubenswrapper[4967]: E1011 04:03:41.152974 4967 reflector.go:158] "Unhandled Error" err="object-\"openshift-controller-manager\"/\"serving-cert\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"serving-cert\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openshift-controller-manager\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.153166 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.153763 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.153936 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.155163 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.162624 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.165326 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-77bc48855d-v7rlc"] Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.172739 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-76ffdc796f-hsm85"] Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.173460 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-76ffdc796f-hsm85" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.179512 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.180008 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.180256 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.180396 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.180566 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.184413 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.184668 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-76ffdc796f-hsm85"] Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.230404 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2189a49-a137-4d13-9f76-92803ed06057-serving-cert\") pod \"route-controller-manager-76ffdc796f-hsm85\" (UID: \"b2189a49-a137-4d13-9f76-92803ed06057\") " pod="openshift-route-controller-manager/route-controller-manager-76ffdc796f-hsm85" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.230443 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e02e6376-7f76-4d1d-9ccb-6349e3b0df0f-proxy-ca-bundles\") pod \"controller-manager-77bc48855d-v7rlc\" (UID: \"e02e6376-7f76-4d1d-9ccb-6349e3b0df0f\") " pod="openshift-controller-manager/controller-manager-77bc48855d-v7rlc" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.230475 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2189a49-a137-4d13-9f76-92803ed06057-config\") pod \"route-controller-manager-76ffdc796f-hsm85\" (UID: \"b2189a49-a137-4d13-9f76-92803ed06057\") " pod="openshift-route-controller-manager/route-controller-manager-76ffdc796f-hsm85" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.230493 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-29tx4\" (UniqueName: \"kubernetes.io/projected/b2189a49-a137-4d13-9f76-92803ed06057-kube-api-access-29tx4\") pod \"route-controller-manager-76ffdc796f-hsm85\" (UID: \"b2189a49-a137-4d13-9f76-92803ed06057\") " pod="openshift-route-controller-manager/route-controller-manager-76ffdc796f-hsm85" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.230514 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b2189a49-a137-4d13-9f76-92803ed06057-client-ca\") pod \"route-controller-manager-76ffdc796f-hsm85\" (UID: \"b2189a49-a137-4d13-9f76-92803ed06057\") " pod="openshift-route-controller-manager/route-controller-manager-76ffdc796f-hsm85" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.230633 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5mnfv\" (UniqueName: \"kubernetes.io/projected/e02e6376-7f76-4d1d-9ccb-6349e3b0df0f-kube-api-access-5mnfv\") pod \"controller-manager-77bc48855d-v7rlc\" (UID: \"e02e6376-7f76-4d1d-9ccb-6349e3b0df0f\") " pod="openshift-controller-manager/controller-manager-77bc48855d-v7rlc" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.230767 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e02e6376-7f76-4d1d-9ccb-6349e3b0df0f-config\") pod \"controller-manager-77bc48855d-v7rlc\" (UID: \"e02e6376-7f76-4d1d-9ccb-6349e3b0df0f\") " pod="openshift-controller-manager/controller-manager-77bc48855d-v7rlc" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.230809 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e02e6376-7f76-4d1d-9ccb-6349e3b0df0f-serving-cert\") pod \"controller-manager-77bc48855d-v7rlc\" (UID: \"e02e6376-7f76-4d1d-9ccb-6349e3b0df0f\") " pod="openshift-controller-manager/controller-manager-77bc48855d-v7rlc" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.230935 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e02e6376-7f76-4d1d-9ccb-6349e3b0df0f-client-ca\") pod \"controller-manager-77bc48855d-v7rlc\" (UID: \"e02e6376-7f76-4d1d-9ccb-6349e3b0df0f\") " pod="openshift-controller-manager/controller-manager-77bc48855d-v7rlc" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.332251 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e02e6376-7f76-4d1d-9ccb-6349e3b0df0f-config\") pod \"controller-manager-77bc48855d-v7rlc\" (UID: \"e02e6376-7f76-4d1d-9ccb-6349e3b0df0f\") " pod="openshift-controller-manager/controller-manager-77bc48855d-v7rlc" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.332296 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e02e6376-7f76-4d1d-9ccb-6349e3b0df0f-serving-cert\") pod \"controller-manager-77bc48855d-v7rlc\" (UID: \"e02e6376-7f76-4d1d-9ccb-6349e3b0df0f\") " pod="openshift-controller-manager/controller-manager-77bc48855d-v7rlc" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.332352 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e02e6376-7f76-4d1d-9ccb-6349e3b0df0f-client-ca\") pod \"controller-manager-77bc48855d-v7rlc\" (UID: \"e02e6376-7f76-4d1d-9ccb-6349e3b0df0f\") " pod="openshift-controller-manager/controller-manager-77bc48855d-v7rlc" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.332483 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2189a49-a137-4d13-9f76-92803ed06057-serving-cert\") pod \"route-controller-manager-76ffdc796f-hsm85\" (UID: \"b2189a49-a137-4d13-9f76-92803ed06057\") " pod="openshift-route-controller-manager/route-controller-manager-76ffdc796f-hsm85" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.332503 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e02e6376-7f76-4d1d-9ccb-6349e3b0df0f-proxy-ca-bundles\") pod \"controller-manager-77bc48855d-v7rlc\" (UID: \"e02e6376-7f76-4d1d-9ccb-6349e3b0df0f\") " pod="openshift-controller-manager/controller-manager-77bc48855d-v7rlc" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.333136 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2189a49-a137-4d13-9f76-92803ed06057-config\") pod \"route-controller-manager-76ffdc796f-hsm85\" (UID: \"b2189a49-a137-4d13-9f76-92803ed06057\") " pod="openshift-route-controller-manager/route-controller-manager-76ffdc796f-hsm85" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.333179 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-29tx4\" (UniqueName: \"kubernetes.io/projected/b2189a49-a137-4d13-9f76-92803ed06057-kube-api-access-29tx4\") pod \"route-controller-manager-76ffdc796f-hsm85\" (UID: \"b2189a49-a137-4d13-9f76-92803ed06057\") " pod="openshift-route-controller-manager/route-controller-manager-76ffdc796f-hsm85" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.333199 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b2189a49-a137-4d13-9f76-92803ed06057-client-ca\") pod \"route-controller-manager-76ffdc796f-hsm85\" (UID: \"b2189a49-a137-4d13-9f76-92803ed06057\") " pod="openshift-route-controller-manager/route-controller-manager-76ffdc796f-hsm85" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.333219 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5mnfv\" (UniqueName: \"kubernetes.io/projected/e02e6376-7f76-4d1d-9ccb-6349e3b0df0f-kube-api-access-5mnfv\") pod \"controller-manager-77bc48855d-v7rlc\" (UID: \"e02e6376-7f76-4d1d-9ccb-6349e3b0df0f\") " pod="openshift-controller-manager/controller-manager-77bc48855d-v7rlc" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.333305 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/e02e6376-7f76-4d1d-9ccb-6349e3b0df0f-client-ca\") pod \"controller-manager-77bc48855d-v7rlc\" (UID: \"e02e6376-7f76-4d1d-9ccb-6349e3b0df0f\") " pod="openshift-controller-manager/controller-manager-77bc48855d-v7rlc" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.333663 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/e02e6376-7f76-4d1d-9ccb-6349e3b0df0f-proxy-ca-bundles\") pod \"controller-manager-77bc48855d-v7rlc\" (UID: \"e02e6376-7f76-4d1d-9ccb-6349e3b0df0f\") " pod="openshift-controller-manager/controller-manager-77bc48855d-v7rlc" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.334195 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2189a49-a137-4d13-9f76-92803ed06057-config\") pod \"route-controller-manager-76ffdc796f-hsm85\" (UID: \"b2189a49-a137-4d13-9f76-92803ed06057\") " pod="openshift-route-controller-manager/route-controller-manager-76ffdc796f-hsm85" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.334690 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e02e6376-7f76-4d1d-9ccb-6349e3b0df0f-config\") pod \"controller-manager-77bc48855d-v7rlc\" (UID: \"e02e6376-7f76-4d1d-9ccb-6349e3b0df0f\") " pod="openshift-controller-manager/controller-manager-77bc48855d-v7rlc" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.337464 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/b2189a49-a137-4d13-9f76-92803ed06057-client-ca\") pod \"route-controller-manager-76ffdc796f-hsm85\" (UID: \"b2189a49-a137-4d13-9f76-92803ed06057\") " pod="openshift-route-controller-manager/route-controller-manager-76ffdc796f-hsm85" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.338194 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b2189a49-a137-4d13-9f76-92803ed06057-serving-cert\") pod \"route-controller-manager-76ffdc796f-hsm85\" (UID: \"b2189a49-a137-4d13-9f76-92803ed06057\") " pod="openshift-route-controller-manager/route-controller-manager-76ffdc796f-hsm85" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.349567 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5mnfv\" (UniqueName: \"kubernetes.io/projected/e02e6376-7f76-4d1d-9ccb-6349e3b0df0f-kube-api-access-5mnfv\") pod \"controller-manager-77bc48855d-v7rlc\" (UID: \"e02e6376-7f76-4d1d-9ccb-6349e3b0df0f\") " pod="openshift-controller-manager/controller-manager-77bc48855d-v7rlc" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.366612 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-29tx4\" (UniqueName: \"kubernetes.io/projected/b2189a49-a137-4d13-9f76-92803ed06057-kube-api-access-29tx4\") pod \"route-controller-manager-76ffdc796f-hsm85\" (UID: \"b2189a49-a137-4d13-9f76-92803ed06057\") " pod="openshift-route-controller-manager/route-controller-manager-76ffdc796f-hsm85" Oct 11 04:03:41 crc kubenswrapper[4967]: I1011 04:03:41.492427 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-76ffdc796f-hsm85" Oct 11 04:03:42 crc kubenswrapper[4967]: I1011 04:03:42.237532 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Oct 11 04:03:42 crc kubenswrapper[4967]: I1011 04:03:42.246156 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e02e6376-7f76-4d1d-9ccb-6349e3b0df0f-serving-cert\") pod \"controller-manager-77bc48855d-v7rlc\" (UID: \"e02e6376-7f76-4d1d-9ccb-6349e3b0df0f\") " pod="openshift-controller-manager/controller-manager-77bc48855d-v7rlc" Oct 11 04:03:42 crc kubenswrapper[4967]: I1011 04:03:42.379704 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-77bc48855d-v7rlc" Oct 11 04:03:43 crc kubenswrapper[4967]: I1011 04:03:43.678218 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-76ffdc796f-hsm85"] Oct 11 04:03:43 crc kubenswrapper[4967]: W1011 04:03:43.703187 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb2189a49_a137_4d13_9f76_92803ed06057.slice/crio-3bfd9ab0164173eb1ae6a78676ccb94329af8d238b8d024a3504a1d1c77afe47 WatchSource:0}: Error finding container 3bfd9ab0164173eb1ae6a78676ccb94329af8d238b8d024a3504a1d1c77afe47: Status 404 returned error can't find the container with id 3bfd9ab0164173eb1ae6a78676ccb94329af8d238b8d024a3504a1d1c77afe47 Oct 11 04:03:43 crc kubenswrapper[4967]: I1011 04:03:43.785060 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-76ffdc796f-hsm85" event={"ID":"b2189a49-a137-4d13-9f76-92803ed06057","Type":"ContainerStarted","Data":"3bfd9ab0164173eb1ae6a78676ccb94329af8d238b8d024a3504a1d1c77afe47"} Oct 11 04:03:43 crc kubenswrapper[4967]: I1011 04:03:43.786248 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-j9lg6" event={"ID":"50110bd6-5363-43c8-b3a8-b4b3762c0c43","Type":"ContainerStarted","Data":"1937c276bf9efd14a6190abf1ee03a2faf92430c68de18d9172f6cc039342633"} Oct 11 04:03:43 crc kubenswrapper[4967]: I1011 04:03:43.787015 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-j9lg6" Oct 11 04:03:43 crc kubenswrapper[4967]: I1011 04:03:43.788859 4967 generic.go:334] "Generic (PLEG): container finished" podID="e9d93993-50a0-468a-a3db-134db9acd6ba" containerID="447d1a55b618f9ec62ae5f80b385f609df07fd9ca0c6ce8df2ffcf3b2705f344" exitCode=0 Oct 11 04:03:43 crc kubenswrapper[4967]: I1011 04:03:43.789122 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-n67vr" event={"ID":"e9d93993-50a0-468a-a3db-134db9acd6ba","Type":"ContainerDied","Data":"447d1a55b618f9ec62ae5f80b385f609df07fd9ca0c6ce8df2ffcf3b2705f344"} Oct 11 04:03:43 crc kubenswrapper[4967]: I1011 04:03:43.822297 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-j9lg6" podStartSLOduration=1.213373626 podStartE2EDuration="8.822278865s" podCreationTimestamp="2025-10-11 04:03:35 +0000 UTC" firstStartedPulling="2025-10-11 04:03:35.969421099 +0000 UTC m=+743.932630032" lastFinishedPulling="2025-10-11 04:03:43.578326338 +0000 UTC m=+751.541535271" observedRunningTime="2025-10-11 04:03:43.802217289 +0000 UTC m=+751.765426222" watchObservedRunningTime="2025-10-11 04:03:43.822278865 +0000 UTC m=+751.785487798" Oct 11 04:03:43 crc kubenswrapper[4967]: I1011 04:03:43.966290 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-77bc48855d-v7rlc"] Oct 11 04:03:43 crc kubenswrapper[4967]: W1011 04:03:43.978324 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode02e6376_7f76_4d1d_9ccb_6349e3b0df0f.slice/crio-2b8d3304dce25f5b7e001ace047707e81925c6962c8382b0b0bf0e8c0001865f WatchSource:0}: Error finding container 2b8d3304dce25f5b7e001ace047707e81925c6962c8382b0b0bf0e8c0001865f: Status 404 returned error can't find the container with id 2b8d3304dce25f5b7e001ace047707e81925c6962c8382b0b0bf0e8c0001865f Oct 11 04:03:44 crc kubenswrapper[4967]: I1011 04:03:44.089706 4967 patch_prober.go:28] interesting pod/machine-config-daemon-sgrv2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:03:44 crc kubenswrapper[4967]: I1011 04:03:44.090269 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:03:44 crc kubenswrapper[4967]: I1011 04:03:44.090422 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" Oct 11 04:03:44 crc kubenswrapper[4967]: I1011 04:03:44.091142 4967 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"fab92459036250055aee13eec5e67ffc42b24499db86755378bf85f3ee4b07d2"} pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 11 04:03:44 crc kubenswrapper[4967]: I1011 04:03:44.091315 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" containerID="cri-o://fab92459036250055aee13eec5e67ffc42b24499db86755378bf85f3ee4b07d2" gracePeriod=600 Oct 11 04:03:44 crc kubenswrapper[4967]: I1011 04:03:44.797731 4967 generic.go:334] "Generic (PLEG): container finished" podID="e9d93993-50a0-468a-a3db-134db9acd6ba" containerID="a7996e29f1b4f1f5cb06c761daeebe72b9d3b751bd19a7040c46a0678cce0ab7" exitCode=0 Oct 11 04:03:44 crc kubenswrapper[4967]: I1011 04:03:44.797798 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-n67vr" event={"ID":"e9d93993-50a0-468a-a3db-134db9acd6ba","Type":"ContainerDied","Data":"a7996e29f1b4f1f5cb06c761daeebe72b9d3b751bd19a7040c46a0678cce0ab7"} Oct 11 04:03:44 crc kubenswrapper[4967]: I1011 04:03:44.800302 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-76ffdc796f-hsm85" event={"ID":"b2189a49-a137-4d13-9f76-92803ed06057","Type":"ContainerStarted","Data":"84864b433eca6de78157ca991a2f75fd67a4dc1e0927ad7983477921fb426773"} Oct 11 04:03:44 crc kubenswrapper[4967]: I1011 04:03:44.802058 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-76ffdc796f-hsm85" Oct 11 04:03:44 crc kubenswrapper[4967]: I1011 04:03:44.805467 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" event={"ID":"97c07678-14be-410c-b61f-498cb49bc960","Type":"ContainerDied","Data":"fab92459036250055aee13eec5e67ffc42b24499db86755378bf85f3ee4b07d2"} Oct 11 04:03:44 crc kubenswrapper[4967]: I1011 04:03:44.805541 4967 scope.go:117] "RemoveContainer" containerID="99f1dbfbe3816d0d177a51d87206e04c54f4faf6dd94db82be3e55ea86cf71f4" Oct 11 04:03:44 crc kubenswrapper[4967]: I1011 04:03:44.805405 4967 generic.go:334] "Generic (PLEG): container finished" podID="97c07678-14be-410c-b61f-498cb49bc960" containerID="fab92459036250055aee13eec5e67ffc42b24499db86755378bf85f3ee4b07d2" exitCode=0 Oct 11 04:03:44 crc kubenswrapper[4967]: I1011 04:03:44.805737 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" event={"ID":"97c07678-14be-410c-b61f-498cb49bc960","Type":"ContainerStarted","Data":"f1443b341ff7c816afffce3e40de74a6868efcc5c30c2eb7be83fc5ffc5860c7"} Oct 11 04:03:44 crc kubenswrapper[4967]: I1011 04:03:44.806059 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-76ffdc796f-hsm85" Oct 11 04:03:44 crc kubenswrapper[4967]: I1011 04:03:44.808019 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-77bc48855d-v7rlc" event={"ID":"e02e6376-7f76-4d1d-9ccb-6349e3b0df0f","Type":"ContainerStarted","Data":"b6532b5410f6b03e929cdf15ad0280a63b201aa12ca14c8f96fa193ecb564d01"} Oct 11 04:03:44 crc kubenswrapper[4967]: I1011 04:03:44.808057 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-77bc48855d-v7rlc" event={"ID":"e02e6376-7f76-4d1d-9ccb-6349e3b0df0f","Type":"ContainerStarted","Data":"2b8d3304dce25f5b7e001ace047707e81925c6962c8382b0b0bf0e8c0001865f"} Oct 11 04:03:44 crc kubenswrapper[4967]: I1011 04:03:44.808340 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-77bc48855d-v7rlc" Oct 11 04:03:44 crc kubenswrapper[4967]: I1011 04:03:44.813272 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-77bc48855d-v7rlc" Oct 11 04:03:44 crc kubenswrapper[4967]: I1011 04:03:44.859358 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-76ffdc796f-hsm85" podStartSLOduration=3.859333586 podStartE2EDuration="3.859333586s" podCreationTimestamp="2025-10-11 04:03:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:03:44.852672441 +0000 UTC m=+752.815881374" watchObservedRunningTime="2025-10-11 04:03:44.859333586 +0000 UTC m=+752.822542519" Oct 11 04:03:44 crc kubenswrapper[4967]: I1011 04:03:44.880175 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-77bc48855d-v7rlc" podStartSLOduration=3.88015444 podStartE2EDuration="3.88015444s" podCreationTimestamp="2025-10-11 04:03:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:03:44.878686316 +0000 UTC m=+752.841895289" watchObservedRunningTime="2025-10-11 04:03:44.88015444 +0000 UTC m=+752.843363393" Oct 11 04:03:45 crc kubenswrapper[4967]: I1011 04:03:45.589022 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/controller-68d546b9d8-rv747" Oct 11 04:03:45 crc kubenswrapper[4967]: I1011 04:03:45.817029 4967 generic.go:334] "Generic (PLEG): container finished" podID="e9d93993-50a0-468a-a3db-134db9acd6ba" containerID="548d19aa1089fc705ecba575b0f6a20120fd12dcef41db2c124fb28313d85b47" exitCode=0 Oct 11 04:03:45 crc kubenswrapper[4967]: I1011 04:03:45.817092 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-n67vr" event={"ID":"e9d93993-50a0-468a-a3db-134db9acd6ba","Type":"ContainerDied","Data":"548d19aa1089fc705ecba575b0f6a20120fd12dcef41db2c124fb28313d85b47"} Oct 11 04:03:46 crc kubenswrapper[4967]: I1011 04:03:46.151426 4967 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Oct 11 04:03:46 crc kubenswrapper[4967]: I1011 04:03:46.162424 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/speaker-np2lm" Oct 11 04:03:46 crc kubenswrapper[4967]: I1011 04:03:46.837118 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-n67vr" event={"ID":"e9d93993-50a0-468a-a3db-134db9acd6ba","Type":"ContainerStarted","Data":"ecf456ee0f78cb71555667aa468eb025bb395745997a6b08b909fc76ce3c1b2c"} Oct 11 04:03:46 crc kubenswrapper[4967]: I1011 04:03:46.839221 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-n67vr" event={"ID":"e9d93993-50a0-468a-a3db-134db9acd6ba","Type":"ContainerStarted","Data":"54864b531c1c0795996aba2c97e0f47e989f588925f34eb0ab0f7a395b3a5b52"} Oct 11 04:03:46 crc kubenswrapper[4967]: I1011 04:03:46.839274 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-n67vr" event={"ID":"e9d93993-50a0-468a-a3db-134db9acd6ba","Type":"ContainerStarted","Data":"b21eb7b636ff0a9124bd2a7ded2e8ea16ea45c4d0a960ecf4bcd1960ad5f6097"} Oct 11 04:03:46 crc kubenswrapper[4967]: I1011 04:03:46.839284 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-n67vr" event={"ID":"e9d93993-50a0-468a-a3db-134db9acd6ba","Type":"ContainerStarted","Data":"3ce1a52afc90434bb508a11af4da42d1be8b14dd552c556d5b4600293c686c8b"} Oct 11 04:03:46 crc kubenswrapper[4967]: I1011 04:03:46.839294 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-n67vr" event={"ID":"e9d93993-50a0-468a-a3db-134db9acd6ba","Type":"ContainerStarted","Data":"2b2917e583dcf09341d494b1a40690a1e8f60ad9ed74c3098b79130de5a51709"} Oct 11 04:03:47 crc kubenswrapper[4967]: I1011 04:03:47.860057 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="metallb-system/frr-k8s-n67vr" event={"ID":"e9d93993-50a0-468a-a3db-134db9acd6ba","Type":"ContainerStarted","Data":"e4005fb8a918df5b86ab348eb0f889b305045f2a4e206aeceddec57bcd5fc581"} Oct 11 04:03:47 crc kubenswrapper[4967]: I1011 04:03:47.860970 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="metallb-system/frr-k8s-n67vr" Oct 11 04:03:47 crc kubenswrapper[4967]: I1011 04:03:47.889057 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="metallb-system/frr-k8s-n67vr" podStartSLOduration=4.950176964 podStartE2EDuration="12.889038078s" podCreationTimestamp="2025-10-11 04:03:35 +0000 UTC" firstStartedPulling="2025-10-11 04:03:35.662428678 +0000 UTC m=+743.625637611" lastFinishedPulling="2025-10-11 04:03:43.601289792 +0000 UTC m=+751.564498725" observedRunningTime="2025-10-11 04:03:47.886367076 +0000 UTC m=+755.849576009" watchObservedRunningTime="2025-10-11 04:03:47.889038078 +0000 UTC m=+755.852247011" Oct 11 04:03:49 crc kubenswrapper[4967]: I1011 04:03:49.214422 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-4795d"] Oct 11 04:03:49 crc kubenswrapper[4967]: I1011 04:03:49.215918 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-4795d" Oct 11 04:03:49 crc kubenswrapper[4967]: I1011 04:03:49.235041 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"kube-root-ca.crt" Oct 11 04:03:49 crc kubenswrapper[4967]: I1011 04:03:49.235284 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack-operators"/"openshift-service-ca.crt" Oct 11 04:03:49 crc kubenswrapper[4967]: I1011 04:03:49.235956 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-4795d"] Oct 11 04:03:49 crc kubenswrapper[4967]: I1011 04:03:49.247460 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c69g4\" (UniqueName: \"kubernetes.io/projected/0111a418-58f6-410e-b326-8a9ca372567f-kube-api-access-c69g4\") pod \"openstack-operator-index-4795d\" (UID: \"0111a418-58f6-410e-b326-8a9ca372567f\") " pod="openstack-operators/openstack-operator-index-4795d" Oct 11 04:03:49 crc kubenswrapper[4967]: I1011 04:03:49.348614 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c69g4\" (UniqueName: \"kubernetes.io/projected/0111a418-58f6-410e-b326-8a9ca372567f-kube-api-access-c69g4\") pod \"openstack-operator-index-4795d\" (UID: \"0111a418-58f6-410e-b326-8a9ca372567f\") " pod="openstack-operators/openstack-operator-index-4795d" Oct 11 04:03:49 crc kubenswrapper[4967]: I1011 04:03:49.366908 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c69g4\" (UniqueName: \"kubernetes.io/projected/0111a418-58f6-410e-b326-8a9ca372567f-kube-api-access-c69g4\") pod \"openstack-operator-index-4795d\" (UID: \"0111a418-58f6-410e-b326-8a9ca372567f\") " pod="openstack-operators/openstack-operator-index-4795d" Oct 11 04:03:49 crc kubenswrapper[4967]: I1011 04:03:49.548536 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-4795d" Oct 11 04:03:49 crc kubenswrapper[4967]: W1011 04:03:49.971693 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0111a418_58f6_410e_b326_8a9ca372567f.slice/crio-b97373f4f8acd117c9433b16b1076cf71ef3a5bce0f37fa8b3df13bfc60ea173 WatchSource:0}: Error finding container b97373f4f8acd117c9433b16b1076cf71ef3a5bce0f37fa8b3df13bfc60ea173: Status 404 returned error can't find the container with id b97373f4f8acd117c9433b16b1076cf71ef3a5bce0f37fa8b3df13bfc60ea173 Oct 11 04:03:49 crc kubenswrapper[4967]: I1011 04:03:49.971906 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-4795d"] Oct 11 04:03:50 crc kubenswrapper[4967]: I1011 04:03:50.500686 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="metallb-system/frr-k8s-n67vr" Oct 11 04:03:50 crc kubenswrapper[4967]: I1011 04:03:50.555271 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="metallb-system/frr-k8s-n67vr" Oct 11 04:03:50 crc kubenswrapper[4967]: I1011 04:03:50.893830 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-4795d" event={"ID":"0111a418-58f6-410e-b326-8a9ca372567f","Type":"ContainerStarted","Data":"b97373f4f8acd117c9433b16b1076cf71ef3a5bce0f37fa8b3df13bfc60ea173"} Oct 11 04:03:52 crc kubenswrapper[4967]: I1011 04:03:52.575691 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-4795d"] Oct 11 04:03:53 crc kubenswrapper[4967]: I1011 04:03:53.181822 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-index-xvswq"] Oct 11 04:03:53 crc kubenswrapper[4967]: I1011 04:03:53.182579 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-xvswq" Oct 11 04:03:53 crc kubenswrapper[4967]: I1011 04:03:53.189101 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-index-dockercfg-ghr9h" Oct 11 04:03:53 crc kubenswrapper[4967]: I1011 04:03:53.193041 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-xvswq"] Oct 11 04:03:53 crc kubenswrapper[4967]: I1011 04:03:53.299493 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h968g\" (UniqueName: \"kubernetes.io/projected/3dafc1e8-5bc6-4b6b-baa4-8c8204d98d31-kube-api-access-h968g\") pod \"openstack-operator-index-xvswq\" (UID: \"3dafc1e8-5bc6-4b6b-baa4-8c8204d98d31\") " pod="openstack-operators/openstack-operator-index-xvswq" Oct 11 04:03:53 crc kubenswrapper[4967]: I1011 04:03:53.400929 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h968g\" (UniqueName: \"kubernetes.io/projected/3dafc1e8-5bc6-4b6b-baa4-8c8204d98d31-kube-api-access-h968g\") pod \"openstack-operator-index-xvswq\" (UID: \"3dafc1e8-5bc6-4b6b-baa4-8c8204d98d31\") " pod="openstack-operators/openstack-operator-index-xvswq" Oct 11 04:03:53 crc kubenswrapper[4967]: I1011 04:03:53.448178 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h968g\" (UniqueName: \"kubernetes.io/projected/3dafc1e8-5bc6-4b6b-baa4-8c8204d98d31-kube-api-access-h968g\") pod \"openstack-operator-index-xvswq\" (UID: \"3dafc1e8-5bc6-4b6b-baa4-8c8204d98d31\") " pod="openstack-operators/openstack-operator-index-xvswq" Oct 11 04:03:53 crc kubenswrapper[4967]: I1011 04:03:53.499769 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-xvswq" Oct 11 04:03:53 crc kubenswrapper[4967]: I1011 04:03:53.917184 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-4795d" event={"ID":"0111a418-58f6-410e-b326-8a9ca372567f","Type":"ContainerStarted","Data":"4f68f156aae2b07baa577cd88f303f7bdb559fbdec6b31340602f35e7877333c"} Oct 11 04:03:53 crc kubenswrapper[4967]: I1011 04:03:53.917351 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack-operators/openstack-operator-index-4795d" podUID="0111a418-58f6-410e-b326-8a9ca372567f" containerName="registry-server" containerID="cri-o://4f68f156aae2b07baa577cd88f303f7bdb559fbdec6b31340602f35e7877333c" gracePeriod=2 Oct 11 04:03:53 crc kubenswrapper[4967]: I1011 04:03:53.966646 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-4795d" podStartSLOduration=1.941788865 podStartE2EDuration="4.966619774s" podCreationTimestamp="2025-10-11 04:03:49 +0000 UTC" firstStartedPulling="2025-10-11 04:03:49.975066987 +0000 UTC m=+757.938275940" lastFinishedPulling="2025-10-11 04:03:52.999897916 +0000 UTC m=+760.963106849" observedRunningTime="2025-10-11 04:03:53.944151272 +0000 UTC m=+761.907360245" watchObservedRunningTime="2025-10-11 04:03:53.966619774 +0000 UTC m=+761.929828747" Oct 11 04:03:53 crc kubenswrapper[4967]: I1011 04:03:53.977568 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-index-xvswq"] Oct 11 04:03:54 crc kubenswrapper[4967]: W1011 04:03:54.002613 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dafc1e8_5bc6_4b6b_baa4_8c8204d98d31.slice/crio-9b4398829bdc1e1a75450a4404a223e9970685a3955ebed8c31d5e63c41f4046 WatchSource:0}: Error finding container 9b4398829bdc1e1a75450a4404a223e9970685a3955ebed8c31d5e63c41f4046: Status 404 returned error can't find the container with id 9b4398829bdc1e1a75450a4404a223e9970685a3955ebed8c31d5e63c41f4046 Oct 11 04:03:54 crc kubenswrapper[4967]: I1011 04:03:54.360263 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-4795d" Oct 11 04:03:54 crc kubenswrapper[4967]: I1011 04:03:54.417332 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c69g4\" (UniqueName: \"kubernetes.io/projected/0111a418-58f6-410e-b326-8a9ca372567f-kube-api-access-c69g4\") pod \"0111a418-58f6-410e-b326-8a9ca372567f\" (UID: \"0111a418-58f6-410e-b326-8a9ca372567f\") " Oct 11 04:03:54 crc kubenswrapper[4967]: I1011 04:03:54.427662 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0111a418-58f6-410e-b326-8a9ca372567f-kube-api-access-c69g4" (OuterVolumeSpecName: "kube-api-access-c69g4") pod "0111a418-58f6-410e-b326-8a9ca372567f" (UID: "0111a418-58f6-410e-b326-8a9ca372567f"). InnerVolumeSpecName "kube-api-access-c69g4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:03:54 crc kubenswrapper[4967]: I1011 04:03:54.518526 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c69g4\" (UniqueName: \"kubernetes.io/projected/0111a418-58f6-410e-b326-8a9ca372567f-kube-api-access-c69g4\") on node \"crc\" DevicePath \"\"" Oct 11 04:03:54 crc kubenswrapper[4967]: I1011 04:03:54.926391 4967 generic.go:334] "Generic (PLEG): container finished" podID="0111a418-58f6-410e-b326-8a9ca372567f" containerID="4f68f156aae2b07baa577cd88f303f7bdb559fbdec6b31340602f35e7877333c" exitCode=0 Oct 11 04:03:54 crc kubenswrapper[4967]: I1011 04:03:54.926483 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-4795d" event={"ID":"0111a418-58f6-410e-b326-8a9ca372567f","Type":"ContainerDied","Data":"4f68f156aae2b07baa577cd88f303f7bdb559fbdec6b31340602f35e7877333c"} Oct 11 04:03:54 crc kubenswrapper[4967]: I1011 04:03:54.926518 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-4795d" event={"ID":"0111a418-58f6-410e-b326-8a9ca372567f","Type":"ContainerDied","Data":"b97373f4f8acd117c9433b16b1076cf71ef3a5bce0f37fa8b3df13bfc60ea173"} Oct 11 04:03:54 crc kubenswrapper[4967]: I1011 04:03:54.926537 4967 scope.go:117] "RemoveContainer" containerID="4f68f156aae2b07baa577cd88f303f7bdb559fbdec6b31340602f35e7877333c" Oct 11 04:03:54 crc kubenswrapper[4967]: I1011 04:03:54.927183 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-index-4795d" Oct 11 04:03:54 crc kubenswrapper[4967]: I1011 04:03:54.928399 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-xvswq" event={"ID":"3dafc1e8-5bc6-4b6b-baa4-8c8204d98d31","Type":"ContainerStarted","Data":"8c1856c37b23e3e8f310e742225c666b7a89bcc2225dd71deecd40c73ea88a22"} Oct 11 04:03:54 crc kubenswrapper[4967]: I1011 04:03:54.928444 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-index-xvswq" event={"ID":"3dafc1e8-5bc6-4b6b-baa4-8c8204d98d31","Type":"ContainerStarted","Data":"9b4398829bdc1e1a75450a4404a223e9970685a3955ebed8c31d5e63c41f4046"} Oct 11 04:03:54 crc kubenswrapper[4967]: I1011 04:03:54.949490 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-index-xvswq" podStartSLOduration=1.894391377 podStartE2EDuration="1.949471526s" podCreationTimestamp="2025-10-11 04:03:53 +0000 UTC" firstStartedPulling="2025-10-11 04:03:54.008837454 +0000 UTC m=+761.972046397" lastFinishedPulling="2025-10-11 04:03:54.063917563 +0000 UTC m=+762.027126546" observedRunningTime="2025-10-11 04:03:54.941651434 +0000 UTC m=+762.904860377" watchObservedRunningTime="2025-10-11 04:03:54.949471526 +0000 UTC m=+762.912680469" Oct 11 04:03:54 crc kubenswrapper[4967]: I1011 04:03:54.966994 4967 scope.go:117] "RemoveContainer" containerID="4f68f156aae2b07baa577cd88f303f7bdb559fbdec6b31340602f35e7877333c" Oct 11 04:03:54 crc kubenswrapper[4967]: E1011 04:03:54.968058 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4f68f156aae2b07baa577cd88f303f7bdb559fbdec6b31340602f35e7877333c\": container with ID starting with 4f68f156aae2b07baa577cd88f303f7bdb559fbdec6b31340602f35e7877333c not found: ID does not exist" containerID="4f68f156aae2b07baa577cd88f303f7bdb559fbdec6b31340602f35e7877333c" Oct 11 04:03:54 crc kubenswrapper[4967]: I1011 04:03:54.968164 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4f68f156aae2b07baa577cd88f303f7bdb559fbdec6b31340602f35e7877333c"} err="failed to get container status \"4f68f156aae2b07baa577cd88f303f7bdb559fbdec6b31340602f35e7877333c\": rpc error: code = NotFound desc = could not find container \"4f68f156aae2b07baa577cd88f303f7bdb559fbdec6b31340602f35e7877333c\": container with ID starting with 4f68f156aae2b07baa577cd88f303f7bdb559fbdec6b31340602f35e7877333c not found: ID does not exist" Oct 11 04:03:54 crc kubenswrapper[4967]: I1011 04:03:54.971809 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack-operators/openstack-operator-index-4795d"] Oct 11 04:03:54 crc kubenswrapper[4967]: I1011 04:03:54.977764 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack-operators/openstack-operator-index-4795d"] Oct 11 04:03:55 crc kubenswrapper[4967]: I1011 04:03:55.486596 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-webhook-server-64bf5d555-j9lg6" Oct 11 04:03:55 crc kubenswrapper[4967]: I1011 04:03:55.505810 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="metallb-system/frr-k8s-n67vr" Oct 11 04:03:56 crc kubenswrapper[4967]: I1011 04:03:56.836566 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0111a418-58f6-410e-b326-8a9ca372567f" path="/var/lib/kubelet/pods/0111a418-58f6-410e-b326-8a9ca372567f/volumes" Oct 11 04:04:01 crc kubenswrapper[4967]: I1011 04:04:01.789792 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-6bgfz"] Oct 11 04:04:01 crc kubenswrapper[4967]: E1011 04:04:01.790700 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0111a418-58f6-410e-b326-8a9ca372567f" containerName="registry-server" Oct 11 04:04:01 crc kubenswrapper[4967]: I1011 04:04:01.790728 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="0111a418-58f6-410e-b326-8a9ca372567f" containerName="registry-server" Oct 11 04:04:01 crc kubenswrapper[4967]: I1011 04:04:01.791032 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="0111a418-58f6-410e-b326-8a9ca372567f" containerName="registry-server" Oct 11 04:04:01 crc kubenswrapper[4967]: I1011 04:04:01.793496 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6bgfz" Oct 11 04:04:01 crc kubenswrapper[4967]: I1011 04:04:01.807927 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6bgfz"] Oct 11 04:04:01 crc kubenswrapper[4967]: I1011 04:04:01.835542 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e91ba0dc-37f9-452c-b02c-649e93873cfd-utilities\") pod \"community-operators-6bgfz\" (UID: \"e91ba0dc-37f9-452c-b02c-649e93873cfd\") " pod="openshift-marketplace/community-operators-6bgfz" Oct 11 04:04:01 crc kubenswrapper[4967]: I1011 04:04:01.836229 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-p2ppp\" (UniqueName: \"kubernetes.io/projected/e91ba0dc-37f9-452c-b02c-649e93873cfd-kube-api-access-p2ppp\") pod \"community-operators-6bgfz\" (UID: \"e91ba0dc-37f9-452c-b02c-649e93873cfd\") " pod="openshift-marketplace/community-operators-6bgfz" Oct 11 04:04:01 crc kubenswrapper[4967]: I1011 04:04:01.836365 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e91ba0dc-37f9-452c-b02c-649e93873cfd-catalog-content\") pod \"community-operators-6bgfz\" (UID: \"e91ba0dc-37f9-452c-b02c-649e93873cfd\") " pod="openshift-marketplace/community-operators-6bgfz" Oct 11 04:04:01 crc kubenswrapper[4967]: I1011 04:04:01.937099 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e91ba0dc-37f9-452c-b02c-649e93873cfd-utilities\") pod \"community-operators-6bgfz\" (UID: \"e91ba0dc-37f9-452c-b02c-649e93873cfd\") " pod="openshift-marketplace/community-operators-6bgfz" Oct 11 04:04:01 crc kubenswrapper[4967]: I1011 04:04:01.937272 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-p2ppp\" (UniqueName: \"kubernetes.io/projected/e91ba0dc-37f9-452c-b02c-649e93873cfd-kube-api-access-p2ppp\") pod \"community-operators-6bgfz\" (UID: \"e91ba0dc-37f9-452c-b02c-649e93873cfd\") " pod="openshift-marketplace/community-operators-6bgfz" Oct 11 04:04:01 crc kubenswrapper[4967]: I1011 04:04:01.937336 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e91ba0dc-37f9-452c-b02c-649e93873cfd-catalog-content\") pod \"community-operators-6bgfz\" (UID: \"e91ba0dc-37f9-452c-b02c-649e93873cfd\") " pod="openshift-marketplace/community-operators-6bgfz" Oct 11 04:04:01 crc kubenswrapper[4967]: I1011 04:04:01.937914 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e91ba0dc-37f9-452c-b02c-649e93873cfd-utilities\") pod \"community-operators-6bgfz\" (UID: \"e91ba0dc-37f9-452c-b02c-649e93873cfd\") " pod="openshift-marketplace/community-operators-6bgfz" Oct 11 04:04:01 crc kubenswrapper[4967]: I1011 04:04:01.938174 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e91ba0dc-37f9-452c-b02c-649e93873cfd-catalog-content\") pod \"community-operators-6bgfz\" (UID: \"e91ba0dc-37f9-452c-b02c-649e93873cfd\") " pod="openshift-marketplace/community-operators-6bgfz" Oct 11 04:04:01 crc kubenswrapper[4967]: I1011 04:04:01.962147 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-p2ppp\" (UniqueName: \"kubernetes.io/projected/e91ba0dc-37f9-452c-b02c-649e93873cfd-kube-api-access-p2ppp\") pod \"community-operators-6bgfz\" (UID: \"e91ba0dc-37f9-452c-b02c-649e93873cfd\") " pod="openshift-marketplace/community-operators-6bgfz" Oct 11 04:04:02 crc kubenswrapper[4967]: I1011 04:04:02.123898 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6bgfz" Oct 11 04:04:02 crc kubenswrapper[4967]: I1011 04:04:02.612838 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-6bgfz"] Oct 11 04:04:02 crc kubenswrapper[4967]: I1011 04:04:02.992342 4967 generic.go:334] "Generic (PLEG): container finished" podID="e91ba0dc-37f9-452c-b02c-649e93873cfd" containerID="7f12f450d3e39d14a9547c5035ede2026223fd28140184a713f265bc54439139" exitCode=0 Oct 11 04:04:02 crc kubenswrapper[4967]: I1011 04:04:02.992490 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6bgfz" event={"ID":"e91ba0dc-37f9-452c-b02c-649e93873cfd","Type":"ContainerDied","Data":"7f12f450d3e39d14a9547c5035ede2026223fd28140184a713f265bc54439139"} Oct 11 04:04:02 crc kubenswrapper[4967]: I1011 04:04:02.992688 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6bgfz" event={"ID":"e91ba0dc-37f9-452c-b02c-649e93873cfd","Type":"ContainerStarted","Data":"23830104aae1c3a2b9d2b2d0398ffbfd1ef87e83a8512b8d5b037f5268775911"} Oct 11 04:04:03 crc kubenswrapper[4967]: I1011 04:04:03.500835 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-index-xvswq" Oct 11 04:04:03 crc kubenswrapper[4967]: I1011 04:04:03.500946 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack-operators/openstack-operator-index-xvswq" Oct 11 04:04:03 crc kubenswrapper[4967]: I1011 04:04:03.574124 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack-operators/openstack-operator-index-xvswq" Oct 11 04:04:04 crc kubenswrapper[4967]: I1011 04:04:04.002130 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6bgfz" event={"ID":"e91ba0dc-37f9-452c-b02c-649e93873cfd","Type":"ContainerStarted","Data":"c1cbb56bb755f3e755ed0434cd22fed91361b5f2135ee2029688cfc0b734e3fd"} Oct 11 04:04:04 crc kubenswrapper[4967]: I1011 04:04:04.030864 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-index-xvswq" Oct 11 04:04:05 crc kubenswrapper[4967]: I1011 04:04:05.013171 4967 generic.go:334] "Generic (PLEG): container finished" podID="e91ba0dc-37f9-452c-b02c-649e93873cfd" containerID="c1cbb56bb755f3e755ed0434cd22fed91361b5f2135ee2029688cfc0b734e3fd" exitCode=0 Oct 11 04:04:05 crc kubenswrapper[4967]: I1011 04:04:05.013257 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6bgfz" event={"ID":"e91ba0dc-37f9-452c-b02c-649e93873cfd","Type":"ContainerDied","Data":"c1cbb56bb755f3e755ed0434cd22fed91361b5f2135ee2029688cfc0b734e3fd"} Oct 11 04:04:06 crc kubenswrapper[4967]: I1011 04:04:06.026683 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6bgfz" event={"ID":"e91ba0dc-37f9-452c-b02c-649e93873cfd","Type":"ContainerStarted","Data":"d4430a897d7127a71ee6e71b5a95d01bbf6716d76e4fcd31f79309917344cd95"} Oct 11 04:04:06 crc kubenswrapper[4967]: I1011 04:04:06.065287 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-6bgfz" podStartSLOduration=2.622003793 podStartE2EDuration="5.065262851s" podCreationTimestamp="2025-10-11 04:04:01 +0000 UTC" firstStartedPulling="2025-10-11 04:04:02.994586308 +0000 UTC m=+770.957795271" lastFinishedPulling="2025-10-11 04:04:05.437845356 +0000 UTC m=+773.401054329" observedRunningTime="2025-10-11 04:04:06.053328274 +0000 UTC m=+774.016537237" watchObservedRunningTime="2025-10-11 04:04:06.065262851 +0000 UTC m=+774.028471814" Oct 11 04:04:11 crc kubenswrapper[4967]: I1011 04:04:11.043925 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk"] Oct 11 04:04:11 crc kubenswrapper[4967]: I1011 04:04:11.046401 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk" Oct 11 04:04:11 crc kubenswrapper[4967]: I1011 04:04:11.049777 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"default-dockercfg-q52wq" Oct 11 04:04:11 crc kubenswrapper[4967]: I1011 04:04:11.057112 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk"] Oct 11 04:04:11 crc kubenswrapper[4967]: I1011 04:04:11.084895 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9adf77e3-238d-4024-99db-ab21bdd51110-util\") pod \"8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk\" (UID: \"9adf77e3-238d-4024-99db-ab21bdd51110\") " pod="openstack-operators/8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk" Oct 11 04:04:11 crc kubenswrapper[4967]: I1011 04:04:11.085579 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tfhqc\" (UniqueName: \"kubernetes.io/projected/9adf77e3-238d-4024-99db-ab21bdd51110-kube-api-access-tfhqc\") pod \"8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk\" (UID: \"9adf77e3-238d-4024-99db-ab21bdd51110\") " pod="openstack-operators/8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk" Oct 11 04:04:11 crc kubenswrapper[4967]: I1011 04:04:11.085645 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9adf77e3-238d-4024-99db-ab21bdd51110-bundle\") pod \"8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk\" (UID: \"9adf77e3-238d-4024-99db-ab21bdd51110\") " pod="openstack-operators/8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk" Oct 11 04:04:11 crc kubenswrapper[4967]: I1011 04:04:11.186242 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9adf77e3-238d-4024-99db-ab21bdd51110-util\") pod \"8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk\" (UID: \"9adf77e3-238d-4024-99db-ab21bdd51110\") " pod="openstack-operators/8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk" Oct 11 04:04:11 crc kubenswrapper[4967]: I1011 04:04:11.186339 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tfhqc\" (UniqueName: \"kubernetes.io/projected/9adf77e3-238d-4024-99db-ab21bdd51110-kube-api-access-tfhqc\") pod \"8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk\" (UID: \"9adf77e3-238d-4024-99db-ab21bdd51110\") " pod="openstack-operators/8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk" Oct 11 04:04:11 crc kubenswrapper[4967]: I1011 04:04:11.186370 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9adf77e3-238d-4024-99db-ab21bdd51110-bundle\") pod \"8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk\" (UID: \"9adf77e3-238d-4024-99db-ab21bdd51110\") " pod="openstack-operators/8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk" Oct 11 04:04:11 crc kubenswrapper[4967]: I1011 04:04:11.186951 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9adf77e3-238d-4024-99db-ab21bdd51110-bundle\") pod \"8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk\" (UID: \"9adf77e3-238d-4024-99db-ab21bdd51110\") " pod="openstack-operators/8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk" Oct 11 04:04:11 crc kubenswrapper[4967]: I1011 04:04:11.187244 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9adf77e3-238d-4024-99db-ab21bdd51110-util\") pod \"8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk\" (UID: \"9adf77e3-238d-4024-99db-ab21bdd51110\") " pod="openstack-operators/8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk" Oct 11 04:04:11 crc kubenswrapper[4967]: I1011 04:04:11.219636 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tfhqc\" (UniqueName: \"kubernetes.io/projected/9adf77e3-238d-4024-99db-ab21bdd51110-kube-api-access-tfhqc\") pod \"8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk\" (UID: \"9adf77e3-238d-4024-99db-ab21bdd51110\") " pod="openstack-operators/8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk" Oct 11 04:04:11 crc kubenswrapper[4967]: I1011 04:04:11.384739 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk" Oct 11 04:04:11 crc kubenswrapper[4967]: I1011 04:04:11.918869 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk"] Oct 11 04:04:12 crc kubenswrapper[4967]: I1011 04:04:12.071404 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk" event={"ID":"9adf77e3-238d-4024-99db-ab21bdd51110","Type":"ContainerStarted","Data":"0769d5a351f53201acbba17bc51694e6be2628309626e703e3e21bcf98f7ac38"} Oct 11 04:04:12 crc kubenswrapper[4967]: I1011 04:04:12.124909 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-6bgfz" Oct 11 04:04:12 crc kubenswrapper[4967]: I1011 04:04:12.124968 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-6bgfz" Oct 11 04:04:12 crc kubenswrapper[4967]: I1011 04:04:12.170668 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-6bgfz" Oct 11 04:04:13 crc kubenswrapper[4967]: I1011 04:04:13.082212 4967 generic.go:334] "Generic (PLEG): container finished" podID="9adf77e3-238d-4024-99db-ab21bdd51110" containerID="5a801dcce82cb4f0d06b88c418b422117d49e4288b98145a8c9e29776046f18b" exitCode=0 Oct 11 04:04:13 crc kubenswrapper[4967]: I1011 04:04:13.082263 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk" event={"ID":"9adf77e3-238d-4024-99db-ab21bdd51110","Type":"ContainerDied","Data":"5a801dcce82cb4f0d06b88c418b422117d49e4288b98145a8c9e29776046f18b"} Oct 11 04:04:13 crc kubenswrapper[4967]: I1011 04:04:13.150489 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-6bgfz" Oct 11 04:04:14 crc kubenswrapper[4967]: I1011 04:04:14.092596 4967 generic.go:334] "Generic (PLEG): container finished" podID="9adf77e3-238d-4024-99db-ab21bdd51110" containerID="cd066004daee56945a53e15b0a11766871c4721d281569daa954be22af2e8db8" exitCode=0 Oct 11 04:04:14 crc kubenswrapper[4967]: I1011 04:04:14.092719 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk" event={"ID":"9adf77e3-238d-4024-99db-ab21bdd51110","Type":"ContainerDied","Data":"cd066004daee56945a53e15b0a11766871c4721d281569daa954be22af2e8db8"} Oct 11 04:04:14 crc kubenswrapper[4967]: I1011 04:04:14.975854 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6bgfz"] Oct 11 04:04:15 crc kubenswrapper[4967]: I1011 04:04:15.105286 4967 generic.go:334] "Generic (PLEG): container finished" podID="9adf77e3-238d-4024-99db-ab21bdd51110" containerID="9268b62e1dc6bbba91f64517bc62cd3c7b616a370e61aa6cbd4f74fd455dc218" exitCode=0 Oct 11 04:04:15 crc kubenswrapper[4967]: I1011 04:04:15.105359 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk" event={"ID":"9adf77e3-238d-4024-99db-ab21bdd51110","Type":"ContainerDied","Data":"9268b62e1dc6bbba91f64517bc62cd3c7b616a370e61aa6cbd4f74fd455dc218"} Oct 11 04:04:15 crc kubenswrapper[4967]: I1011 04:04:15.105560 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-6bgfz" podUID="e91ba0dc-37f9-452c-b02c-649e93873cfd" containerName="registry-server" containerID="cri-o://d4430a897d7127a71ee6e71b5a95d01bbf6716d76e4fcd31f79309917344cd95" gracePeriod=2 Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.112533 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6bgfz" Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.118451 4967 generic.go:334] "Generic (PLEG): container finished" podID="e91ba0dc-37f9-452c-b02c-649e93873cfd" containerID="d4430a897d7127a71ee6e71b5a95d01bbf6716d76e4fcd31f79309917344cd95" exitCode=0 Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.118534 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-6bgfz" Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.118563 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6bgfz" event={"ID":"e91ba0dc-37f9-452c-b02c-649e93873cfd","Type":"ContainerDied","Data":"d4430a897d7127a71ee6e71b5a95d01bbf6716d76e4fcd31f79309917344cd95"} Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.118638 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-6bgfz" event={"ID":"e91ba0dc-37f9-452c-b02c-649e93873cfd","Type":"ContainerDied","Data":"23830104aae1c3a2b9d2b2d0398ffbfd1ef87e83a8512b8d5b037f5268775911"} Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.118666 4967 scope.go:117] "RemoveContainer" containerID="d4430a897d7127a71ee6e71b5a95d01bbf6716d76e4fcd31f79309917344cd95" Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.167789 4967 scope.go:117] "RemoveContainer" containerID="c1cbb56bb755f3e755ed0434cd22fed91361b5f2135ee2029688cfc0b734e3fd" Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.192107 4967 scope.go:117] "RemoveContainer" containerID="7f12f450d3e39d14a9547c5035ede2026223fd28140184a713f265bc54439139" Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.219631 4967 scope.go:117] "RemoveContainer" containerID="d4430a897d7127a71ee6e71b5a95d01bbf6716d76e4fcd31f79309917344cd95" Oct 11 04:04:16 crc kubenswrapper[4967]: E1011 04:04:16.220855 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4430a897d7127a71ee6e71b5a95d01bbf6716d76e4fcd31f79309917344cd95\": container with ID starting with d4430a897d7127a71ee6e71b5a95d01bbf6716d76e4fcd31f79309917344cd95 not found: ID does not exist" containerID="d4430a897d7127a71ee6e71b5a95d01bbf6716d76e4fcd31f79309917344cd95" Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.220900 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4430a897d7127a71ee6e71b5a95d01bbf6716d76e4fcd31f79309917344cd95"} err="failed to get container status \"d4430a897d7127a71ee6e71b5a95d01bbf6716d76e4fcd31f79309917344cd95\": rpc error: code = NotFound desc = could not find container \"d4430a897d7127a71ee6e71b5a95d01bbf6716d76e4fcd31f79309917344cd95\": container with ID starting with d4430a897d7127a71ee6e71b5a95d01bbf6716d76e4fcd31f79309917344cd95 not found: ID does not exist" Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.220927 4967 scope.go:117] "RemoveContainer" containerID="c1cbb56bb755f3e755ed0434cd22fed91361b5f2135ee2029688cfc0b734e3fd" Oct 11 04:04:16 crc kubenswrapper[4967]: E1011 04:04:16.221541 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c1cbb56bb755f3e755ed0434cd22fed91361b5f2135ee2029688cfc0b734e3fd\": container with ID starting with c1cbb56bb755f3e755ed0434cd22fed91361b5f2135ee2029688cfc0b734e3fd not found: ID does not exist" containerID="c1cbb56bb755f3e755ed0434cd22fed91361b5f2135ee2029688cfc0b734e3fd" Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.221615 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c1cbb56bb755f3e755ed0434cd22fed91361b5f2135ee2029688cfc0b734e3fd"} err="failed to get container status \"c1cbb56bb755f3e755ed0434cd22fed91361b5f2135ee2029688cfc0b734e3fd\": rpc error: code = NotFound desc = could not find container \"c1cbb56bb755f3e755ed0434cd22fed91361b5f2135ee2029688cfc0b734e3fd\": container with ID starting with c1cbb56bb755f3e755ed0434cd22fed91361b5f2135ee2029688cfc0b734e3fd not found: ID does not exist" Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.221650 4967 scope.go:117] "RemoveContainer" containerID="7f12f450d3e39d14a9547c5035ede2026223fd28140184a713f265bc54439139" Oct 11 04:04:16 crc kubenswrapper[4967]: E1011 04:04:16.222322 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f12f450d3e39d14a9547c5035ede2026223fd28140184a713f265bc54439139\": container with ID starting with 7f12f450d3e39d14a9547c5035ede2026223fd28140184a713f265bc54439139 not found: ID does not exist" containerID="7f12f450d3e39d14a9547c5035ede2026223fd28140184a713f265bc54439139" Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.222355 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f12f450d3e39d14a9547c5035ede2026223fd28140184a713f265bc54439139"} err="failed to get container status \"7f12f450d3e39d14a9547c5035ede2026223fd28140184a713f265bc54439139\": rpc error: code = NotFound desc = could not find container \"7f12f450d3e39d14a9547c5035ede2026223fd28140184a713f265bc54439139\": container with ID starting with 7f12f450d3e39d14a9547c5035ede2026223fd28140184a713f265bc54439139 not found: ID does not exist" Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.269238 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-p2ppp\" (UniqueName: \"kubernetes.io/projected/e91ba0dc-37f9-452c-b02c-649e93873cfd-kube-api-access-p2ppp\") pod \"e91ba0dc-37f9-452c-b02c-649e93873cfd\" (UID: \"e91ba0dc-37f9-452c-b02c-649e93873cfd\") " Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.269381 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e91ba0dc-37f9-452c-b02c-649e93873cfd-utilities\") pod \"e91ba0dc-37f9-452c-b02c-649e93873cfd\" (UID: \"e91ba0dc-37f9-452c-b02c-649e93873cfd\") " Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.269439 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e91ba0dc-37f9-452c-b02c-649e93873cfd-catalog-content\") pod \"e91ba0dc-37f9-452c-b02c-649e93873cfd\" (UID: \"e91ba0dc-37f9-452c-b02c-649e93873cfd\") " Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.270558 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e91ba0dc-37f9-452c-b02c-649e93873cfd-utilities" (OuterVolumeSpecName: "utilities") pod "e91ba0dc-37f9-452c-b02c-649e93873cfd" (UID: "e91ba0dc-37f9-452c-b02c-649e93873cfd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.290261 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e91ba0dc-37f9-452c-b02c-649e93873cfd-kube-api-access-p2ppp" (OuterVolumeSpecName: "kube-api-access-p2ppp") pod "e91ba0dc-37f9-452c-b02c-649e93873cfd" (UID: "e91ba0dc-37f9-452c-b02c-649e93873cfd"). InnerVolumeSpecName "kube-api-access-p2ppp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.313430 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e91ba0dc-37f9-452c-b02c-649e93873cfd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e91ba0dc-37f9-452c-b02c-649e93873cfd" (UID: "e91ba0dc-37f9-452c-b02c-649e93873cfd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.371277 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e91ba0dc-37f9-452c-b02c-649e93873cfd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.371309 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-p2ppp\" (UniqueName: \"kubernetes.io/projected/e91ba0dc-37f9-452c-b02c-649e93873cfd-kube-api-access-p2ppp\") on node \"crc\" DevicePath \"\"" Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.371324 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e91ba0dc-37f9-452c-b02c-649e93873cfd-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.483225 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk" Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.499179 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-6bgfz"] Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.514167 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-6bgfz"] Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.675634 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9adf77e3-238d-4024-99db-ab21bdd51110-util\") pod \"9adf77e3-238d-4024-99db-ab21bdd51110\" (UID: \"9adf77e3-238d-4024-99db-ab21bdd51110\") " Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.675905 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tfhqc\" (UniqueName: \"kubernetes.io/projected/9adf77e3-238d-4024-99db-ab21bdd51110-kube-api-access-tfhqc\") pod \"9adf77e3-238d-4024-99db-ab21bdd51110\" (UID: \"9adf77e3-238d-4024-99db-ab21bdd51110\") " Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.675963 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9adf77e3-238d-4024-99db-ab21bdd51110-bundle\") pod \"9adf77e3-238d-4024-99db-ab21bdd51110\" (UID: \"9adf77e3-238d-4024-99db-ab21bdd51110\") " Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.677149 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9adf77e3-238d-4024-99db-ab21bdd51110-bundle" (OuterVolumeSpecName: "bundle") pod "9adf77e3-238d-4024-99db-ab21bdd51110" (UID: "9adf77e3-238d-4024-99db-ab21bdd51110"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.682259 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9adf77e3-238d-4024-99db-ab21bdd51110-kube-api-access-tfhqc" (OuterVolumeSpecName: "kube-api-access-tfhqc") pod "9adf77e3-238d-4024-99db-ab21bdd51110" (UID: "9adf77e3-238d-4024-99db-ab21bdd51110"). InnerVolumeSpecName "kube-api-access-tfhqc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.705749 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9adf77e3-238d-4024-99db-ab21bdd51110-util" (OuterVolumeSpecName: "util") pod "9adf77e3-238d-4024-99db-ab21bdd51110" (UID: "9adf77e3-238d-4024-99db-ab21bdd51110"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.777733 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tfhqc\" (UniqueName: \"kubernetes.io/projected/9adf77e3-238d-4024-99db-ab21bdd51110-kube-api-access-tfhqc\") on node \"crc\" DevicePath \"\"" Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.777795 4967 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/9adf77e3-238d-4024-99db-ab21bdd51110-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.777814 4967 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/9adf77e3-238d-4024-99db-ab21bdd51110-util\") on node \"crc\" DevicePath \"\"" Oct 11 04:04:16 crc kubenswrapper[4967]: I1011 04:04:16.830833 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e91ba0dc-37f9-452c-b02c-649e93873cfd" path="/var/lib/kubelet/pods/e91ba0dc-37f9-452c-b02c-649e93873cfd/volumes" Oct 11 04:04:17 crc kubenswrapper[4967]: I1011 04:04:17.137296 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk" event={"ID":"9adf77e3-238d-4024-99db-ab21bdd51110","Type":"ContainerDied","Data":"0769d5a351f53201acbba17bc51694e6be2628309626e703e3e21bcf98f7ac38"} Oct 11 04:04:17 crc kubenswrapper[4967]: I1011 04:04:17.137349 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0769d5a351f53201acbba17bc51694e6be2628309626e703e3e21bcf98f7ac38" Oct 11 04:04:17 crc kubenswrapper[4967]: I1011 04:04:17.137407 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack-operators/8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk" Oct 11 04:04:20 crc kubenswrapper[4967]: I1011 04:04:20.138429 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-operator-84859cd6c-hlmz6"] Oct 11 04:04:20 crc kubenswrapper[4967]: E1011 04:04:20.139033 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e91ba0dc-37f9-452c-b02c-649e93873cfd" containerName="extract-content" Oct 11 04:04:20 crc kubenswrapper[4967]: I1011 04:04:20.139051 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="e91ba0dc-37f9-452c-b02c-649e93873cfd" containerName="extract-content" Oct 11 04:04:20 crc kubenswrapper[4967]: E1011 04:04:20.139100 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9adf77e3-238d-4024-99db-ab21bdd51110" containerName="util" Oct 11 04:04:20 crc kubenswrapper[4967]: I1011 04:04:20.139110 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="9adf77e3-238d-4024-99db-ab21bdd51110" containerName="util" Oct 11 04:04:20 crc kubenswrapper[4967]: E1011 04:04:20.139126 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9adf77e3-238d-4024-99db-ab21bdd51110" containerName="extract" Oct 11 04:04:20 crc kubenswrapper[4967]: I1011 04:04:20.139135 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="9adf77e3-238d-4024-99db-ab21bdd51110" containerName="extract" Oct 11 04:04:20 crc kubenswrapper[4967]: E1011 04:04:20.139149 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9adf77e3-238d-4024-99db-ab21bdd51110" containerName="pull" Oct 11 04:04:20 crc kubenswrapper[4967]: I1011 04:04:20.139156 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="9adf77e3-238d-4024-99db-ab21bdd51110" containerName="pull" Oct 11 04:04:20 crc kubenswrapper[4967]: E1011 04:04:20.139176 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e91ba0dc-37f9-452c-b02c-649e93873cfd" containerName="registry-server" Oct 11 04:04:20 crc kubenswrapper[4967]: I1011 04:04:20.139184 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="e91ba0dc-37f9-452c-b02c-649e93873cfd" containerName="registry-server" Oct 11 04:04:20 crc kubenswrapper[4967]: E1011 04:04:20.139197 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e91ba0dc-37f9-452c-b02c-649e93873cfd" containerName="extract-utilities" Oct 11 04:04:20 crc kubenswrapper[4967]: I1011 04:04:20.139207 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="e91ba0dc-37f9-452c-b02c-649e93873cfd" containerName="extract-utilities" Oct 11 04:04:20 crc kubenswrapper[4967]: I1011 04:04:20.139344 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="e91ba0dc-37f9-452c-b02c-649e93873cfd" containerName="registry-server" Oct 11 04:04:20 crc kubenswrapper[4967]: I1011 04:04:20.139365 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="9adf77e3-238d-4024-99db-ab21bdd51110" containerName="extract" Oct 11 04:04:20 crc kubenswrapper[4967]: I1011 04:04:20.140116 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-84859cd6c-hlmz6" Oct 11 04:04:20 crc kubenswrapper[4967]: I1011 04:04:20.154596 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-operator-dockercfg-ppzsm" Oct 11 04:04:20 crc kubenswrapper[4967]: I1011 04:04:20.162314 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-84859cd6c-hlmz6"] Oct 11 04:04:20 crc kubenswrapper[4967]: I1011 04:04:20.325228 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tflnx\" (UniqueName: \"kubernetes.io/projected/e36d39c8-9b9a-4564-b779-0c038ba68dfb-kube-api-access-tflnx\") pod \"openstack-operator-controller-operator-84859cd6c-hlmz6\" (UID: \"e36d39c8-9b9a-4564-b779-0c038ba68dfb\") " pod="openstack-operators/openstack-operator-controller-operator-84859cd6c-hlmz6" Oct 11 04:04:20 crc kubenswrapper[4967]: I1011 04:04:20.426711 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tflnx\" (UniqueName: \"kubernetes.io/projected/e36d39c8-9b9a-4564-b779-0c038ba68dfb-kube-api-access-tflnx\") pod \"openstack-operator-controller-operator-84859cd6c-hlmz6\" (UID: \"e36d39c8-9b9a-4564-b779-0c038ba68dfb\") " pod="openstack-operators/openstack-operator-controller-operator-84859cd6c-hlmz6" Oct 11 04:04:20 crc kubenswrapper[4967]: I1011 04:04:20.446004 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tflnx\" (UniqueName: \"kubernetes.io/projected/e36d39c8-9b9a-4564-b779-0c038ba68dfb-kube-api-access-tflnx\") pod \"openstack-operator-controller-operator-84859cd6c-hlmz6\" (UID: \"e36d39c8-9b9a-4564-b779-0c038ba68dfb\") " pod="openstack-operators/openstack-operator-controller-operator-84859cd6c-hlmz6" Oct 11 04:04:20 crc kubenswrapper[4967]: I1011 04:04:20.456296 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-operator-84859cd6c-hlmz6" Oct 11 04:04:20 crc kubenswrapper[4967]: I1011 04:04:20.931226 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-operator-84859cd6c-hlmz6"] Oct 11 04:04:21 crc kubenswrapper[4967]: I1011 04:04:21.164936 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-84859cd6c-hlmz6" event={"ID":"e36d39c8-9b9a-4564-b779-0c038ba68dfb","Type":"ContainerStarted","Data":"d41d5d3f473bcdd3525498c72f292f0f6194f6467685e8f1f9bd4542266ecc08"} Oct 11 04:04:25 crc kubenswrapper[4967]: I1011 04:04:25.202421 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-84859cd6c-hlmz6" event={"ID":"e36d39c8-9b9a-4564-b779-0c038ba68dfb","Type":"ContainerStarted","Data":"845c5b4af6c1505162c045fd7e0e297145bcdac5966286efdf0eb156d93c6bed"} Oct 11 04:04:27 crc kubenswrapper[4967]: I1011 04:04:27.232241 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-operator-84859cd6c-hlmz6" event={"ID":"e36d39c8-9b9a-4564-b779-0c038ba68dfb","Type":"ContainerStarted","Data":"7e9c2e274951888d3998cc96264971492752e63970720daaf1668cc6a20c8e8e"} Oct 11 04:04:27 crc kubenswrapper[4967]: I1011 04:04:27.233190 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-operator-84859cd6c-hlmz6" Oct 11 04:04:27 crc kubenswrapper[4967]: I1011 04:04:27.270525 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-operator-84859cd6c-hlmz6" podStartSLOduration=1.383255426 podStartE2EDuration="7.270501456s" podCreationTimestamp="2025-10-11 04:04:20 +0000 UTC" firstStartedPulling="2025-10-11 04:04:20.948511221 +0000 UTC m=+788.911720154" lastFinishedPulling="2025-10-11 04:04:26.835757251 +0000 UTC m=+794.798966184" observedRunningTime="2025-10-11 04:04:27.268053598 +0000 UTC m=+795.231262551" watchObservedRunningTime="2025-10-11 04:04:27.270501456 +0000 UTC m=+795.233710399" Oct 11 04:04:30 crc kubenswrapper[4967]: I1011 04:04:30.462959 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-operator-84859cd6c-hlmz6" Oct 11 04:04:38 crc kubenswrapper[4967]: I1011 04:04:38.740218 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-8nwxv"] Oct 11 04:04:38 crc kubenswrapper[4967]: I1011 04:04:38.742020 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8nwxv" Oct 11 04:04:38 crc kubenswrapper[4967]: I1011 04:04:38.748152 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8nwxv"] Oct 11 04:04:38 crc kubenswrapper[4967]: I1011 04:04:38.878574 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8487ea91-dfea-423a-8c19-358a235b9eff-catalog-content\") pod \"redhat-operators-8nwxv\" (UID: \"8487ea91-dfea-423a-8c19-358a235b9eff\") " pod="openshift-marketplace/redhat-operators-8nwxv" Oct 11 04:04:38 crc kubenswrapper[4967]: I1011 04:04:38.878941 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8487ea91-dfea-423a-8c19-358a235b9eff-utilities\") pod \"redhat-operators-8nwxv\" (UID: \"8487ea91-dfea-423a-8c19-358a235b9eff\") " pod="openshift-marketplace/redhat-operators-8nwxv" Oct 11 04:04:38 crc kubenswrapper[4967]: I1011 04:04:38.879132 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xpvw8\" (UniqueName: \"kubernetes.io/projected/8487ea91-dfea-423a-8c19-358a235b9eff-kube-api-access-xpvw8\") pod \"redhat-operators-8nwxv\" (UID: \"8487ea91-dfea-423a-8c19-358a235b9eff\") " pod="openshift-marketplace/redhat-operators-8nwxv" Oct 11 04:04:38 crc kubenswrapper[4967]: I1011 04:04:38.980958 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8487ea91-dfea-423a-8c19-358a235b9eff-catalog-content\") pod \"redhat-operators-8nwxv\" (UID: \"8487ea91-dfea-423a-8c19-358a235b9eff\") " pod="openshift-marketplace/redhat-operators-8nwxv" Oct 11 04:04:38 crc kubenswrapper[4967]: I1011 04:04:38.981064 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8487ea91-dfea-423a-8c19-358a235b9eff-utilities\") pod \"redhat-operators-8nwxv\" (UID: \"8487ea91-dfea-423a-8c19-358a235b9eff\") " pod="openshift-marketplace/redhat-operators-8nwxv" Oct 11 04:04:38 crc kubenswrapper[4967]: I1011 04:04:38.981113 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xpvw8\" (UniqueName: \"kubernetes.io/projected/8487ea91-dfea-423a-8c19-358a235b9eff-kube-api-access-xpvw8\") pod \"redhat-operators-8nwxv\" (UID: \"8487ea91-dfea-423a-8c19-358a235b9eff\") " pod="openshift-marketplace/redhat-operators-8nwxv" Oct 11 04:04:38 crc kubenswrapper[4967]: I1011 04:04:38.981644 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8487ea91-dfea-423a-8c19-358a235b9eff-catalog-content\") pod \"redhat-operators-8nwxv\" (UID: \"8487ea91-dfea-423a-8c19-358a235b9eff\") " pod="openshift-marketplace/redhat-operators-8nwxv" Oct 11 04:04:38 crc kubenswrapper[4967]: I1011 04:04:38.981731 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8487ea91-dfea-423a-8c19-358a235b9eff-utilities\") pod \"redhat-operators-8nwxv\" (UID: \"8487ea91-dfea-423a-8c19-358a235b9eff\") " pod="openshift-marketplace/redhat-operators-8nwxv" Oct 11 04:04:39 crc kubenswrapper[4967]: I1011 04:04:39.000469 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xpvw8\" (UniqueName: \"kubernetes.io/projected/8487ea91-dfea-423a-8c19-358a235b9eff-kube-api-access-xpvw8\") pod \"redhat-operators-8nwxv\" (UID: \"8487ea91-dfea-423a-8c19-358a235b9eff\") " pod="openshift-marketplace/redhat-operators-8nwxv" Oct 11 04:04:39 crc kubenswrapper[4967]: I1011 04:04:39.064544 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8nwxv" Oct 11 04:04:39 crc kubenswrapper[4967]: I1011 04:04:39.514089 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-8nwxv"] Oct 11 04:04:40 crc kubenswrapper[4967]: I1011 04:04:40.325938 4967 generic.go:334] "Generic (PLEG): container finished" podID="8487ea91-dfea-423a-8c19-358a235b9eff" containerID="275e549b6a7b3950e3c225c5343d97b681645718fe8cbdbcfb1f4325fc5b6b27" exitCode=0 Oct 11 04:04:40 crc kubenswrapper[4967]: I1011 04:04:40.326007 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8nwxv" event={"ID":"8487ea91-dfea-423a-8c19-358a235b9eff","Type":"ContainerDied","Data":"275e549b6a7b3950e3c225c5343d97b681645718fe8cbdbcfb1f4325fc5b6b27"} Oct 11 04:04:40 crc kubenswrapper[4967]: I1011 04:04:40.326198 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8nwxv" event={"ID":"8487ea91-dfea-423a-8c19-358a235b9eff","Type":"ContainerStarted","Data":"c067624ff054a45efe5dcf20c045d1ed3750a1a91365a4a142a8fc8994993538"} Oct 11 04:04:41 crc kubenswrapper[4967]: I1011 04:04:41.333903 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8nwxv" event={"ID":"8487ea91-dfea-423a-8c19-358a235b9eff","Type":"ContainerStarted","Data":"1c4cb8afea13c57b22e73f4481c68854e398403effb4fc446167bb32a8fd3280"} Oct 11 04:04:42 crc kubenswrapper[4967]: I1011 04:04:42.340410 4967 generic.go:334] "Generic (PLEG): container finished" podID="8487ea91-dfea-423a-8c19-358a235b9eff" containerID="1c4cb8afea13c57b22e73f4481c68854e398403effb4fc446167bb32a8fd3280" exitCode=0 Oct 11 04:04:42 crc kubenswrapper[4967]: I1011 04:04:42.340462 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8nwxv" event={"ID":"8487ea91-dfea-423a-8c19-358a235b9eff","Type":"ContainerDied","Data":"1c4cb8afea13c57b22e73f4481c68854e398403effb4fc446167bb32a8fd3280"} Oct 11 04:04:43 crc kubenswrapper[4967]: I1011 04:04:43.348183 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8nwxv" event={"ID":"8487ea91-dfea-423a-8c19-358a235b9eff","Type":"ContainerStarted","Data":"64e21d0eed5b8671f1f847782f3177afa710a544e25e1713597cfa9e2ce9d877"} Oct 11 04:04:43 crc kubenswrapper[4967]: I1011 04:04:43.367386 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-8nwxv" podStartSLOduration=2.745697782 podStartE2EDuration="5.367371665s" podCreationTimestamp="2025-10-11 04:04:38 +0000 UTC" firstStartedPulling="2025-10-11 04:04:40.327728904 +0000 UTC m=+808.290937837" lastFinishedPulling="2025-10-11 04:04:42.949402797 +0000 UTC m=+810.912611720" observedRunningTime="2025-10-11 04:04:43.364821516 +0000 UTC m=+811.328030449" watchObservedRunningTime="2025-10-11 04:04:43.367371665 +0000 UTC m=+811.330580598" Oct 11 04:04:45 crc kubenswrapper[4967]: I1011 04:04:45.310161 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-z9scd"] Oct 11 04:04:45 crc kubenswrapper[4967]: I1011 04:04:45.311996 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z9scd" Oct 11 04:04:45 crc kubenswrapper[4967]: I1011 04:04:45.322984 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z9scd"] Oct 11 04:04:45 crc kubenswrapper[4967]: I1011 04:04:45.469852 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7de04fa6-4dd6-4254-bde0-b83b920be0fa-catalog-content\") pod \"redhat-marketplace-z9scd\" (UID: \"7de04fa6-4dd6-4254-bde0-b83b920be0fa\") " pod="openshift-marketplace/redhat-marketplace-z9scd" Oct 11 04:04:45 crc kubenswrapper[4967]: I1011 04:04:45.469900 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96b89\" (UniqueName: \"kubernetes.io/projected/7de04fa6-4dd6-4254-bde0-b83b920be0fa-kube-api-access-96b89\") pod \"redhat-marketplace-z9scd\" (UID: \"7de04fa6-4dd6-4254-bde0-b83b920be0fa\") " pod="openshift-marketplace/redhat-marketplace-z9scd" Oct 11 04:04:45 crc kubenswrapper[4967]: I1011 04:04:45.469940 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7de04fa6-4dd6-4254-bde0-b83b920be0fa-utilities\") pod \"redhat-marketplace-z9scd\" (UID: \"7de04fa6-4dd6-4254-bde0-b83b920be0fa\") " pod="openshift-marketplace/redhat-marketplace-z9scd" Oct 11 04:04:45 crc kubenswrapper[4967]: I1011 04:04:45.570918 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7de04fa6-4dd6-4254-bde0-b83b920be0fa-catalog-content\") pod \"redhat-marketplace-z9scd\" (UID: \"7de04fa6-4dd6-4254-bde0-b83b920be0fa\") " pod="openshift-marketplace/redhat-marketplace-z9scd" Oct 11 04:04:45 crc kubenswrapper[4967]: I1011 04:04:45.570976 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96b89\" (UniqueName: \"kubernetes.io/projected/7de04fa6-4dd6-4254-bde0-b83b920be0fa-kube-api-access-96b89\") pod \"redhat-marketplace-z9scd\" (UID: \"7de04fa6-4dd6-4254-bde0-b83b920be0fa\") " pod="openshift-marketplace/redhat-marketplace-z9scd" Oct 11 04:04:45 crc kubenswrapper[4967]: I1011 04:04:45.571019 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7de04fa6-4dd6-4254-bde0-b83b920be0fa-utilities\") pod \"redhat-marketplace-z9scd\" (UID: \"7de04fa6-4dd6-4254-bde0-b83b920be0fa\") " pod="openshift-marketplace/redhat-marketplace-z9scd" Oct 11 04:04:45 crc kubenswrapper[4967]: I1011 04:04:45.571514 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7de04fa6-4dd6-4254-bde0-b83b920be0fa-catalog-content\") pod \"redhat-marketplace-z9scd\" (UID: \"7de04fa6-4dd6-4254-bde0-b83b920be0fa\") " pod="openshift-marketplace/redhat-marketplace-z9scd" Oct 11 04:04:45 crc kubenswrapper[4967]: I1011 04:04:45.571530 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7de04fa6-4dd6-4254-bde0-b83b920be0fa-utilities\") pod \"redhat-marketplace-z9scd\" (UID: \"7de04fa6-4dd6-4254-bde0-b83b920be0fa\") " pod="openshift-marketplace/redhat-marketplace-z9scd" Oct 11 04:04:45 crc kubenswrapper[4967]: I1011 04:04:45.600051 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96b89\" (UniqueName: \"kubernetes.io/projected/7de04fa6-4dd6-4254-bde0-b83b920be0fa-kube-api-access-96b89\") pod \"redhat-marketplace-z9scd\" (UID: \"7de04fa6-4dd6-4254-bde0-b83b920be0fa\") " pod="openshift-marketplace/redhat-marketplace-z9scd" Oct 11 04:04:45 crc kubenswrapper[4967]: I1011 04:04:45.626719 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z9scd" Oct 11 04:04:46 crc kubenswrapper[4967]: I1011 04:04:46.053094 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-z9scd"] Oct 11 04:04:46 crc kubenswrapper[4967]: I1011 04:04:46.388925 4967 generic.go:334] "Generic (PLEG): container finished" podID="7de04fa6-4dd6-4254-bde0-b83b920be0fa" containerID="83962bf9b636c5e9588f7254c872df4d327b6b03b40c0521f28e97227e48a89d" exitCode=0 Oct 11 04:04:46 crc kubenswrapper[4967]: I1011 04:04:46.388986 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z9scd" event={"ID":"7de04fa6-4dd6-4254-bde0-b83b920be0fa","Type":"ContainerDied","Data":"83962bf9b636c5e9588f7254c872df4d327b6b03b40c0521f28e97227e48a89d"} Oct 11 04:04:46 crc kubenswrapper[4967]: I1011 04:04:46.389235 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z9scd" event={"ID":"7de04fa6-4dd6-4254-bde0-b83b920be0fa","Type":"ContainerStarted","Data":"148bab79652295b728efd895d90bbd1c9011a5108116cdd9cf94d7f264961b3a"} Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.443993 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f84fcdbb-xz92m"] Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.445125 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-xz92m" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.446793 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"barbican-operator-controller-manager-dockercfg-dkj99" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.453214 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/cinder-operator-controller-manager-59cdc64769-5glgp"] Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.454812 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-5glgp" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.456391 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f84fcdbb-xz92m"] Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.466306 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"cinder-operator-controller-manager-dockercfg-zsbdd" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.470491 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/designate-operator-controller-manager-687df44cdb-9gn9l"] Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.471681 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-9gn9l" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.474230 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"designate-operator-controller-manager-dockercfg-77cl9" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.475440 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-59cdc64769-5glgp"] Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.518719 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/glance-operator-controller-manager-7bb46cd7d-zt44b"] Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.520181 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-zt44b" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.523208 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"glance-operator-controller-manager-dockercfg-4tqv7" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.532027 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-687df44cdb-9gn9l"] Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.545491 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7bb46cd7d-zt44b"] Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.579434 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/heat-operator-controller-manager-6d9967f8dd-ww7nt"] Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.580452 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-ww7nt" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.581589 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-6d9967f8dd-ww7nt"] Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.583614 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"heat-operator-controller-manager-dockercfg-gn2tr" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.596637 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z7g9q\" (UniqueName: \"kubernetes.io/projected/2a2805ea-7d92-42e4-89cb-9b2672f72fc6-kube-api-access-z7g9q\") pod \"designate-operator-controller-manager-687df44cdb-9gn9l\" (UID: \"2a2805ea-7d92-42e4-89cb-9b2672f72fc6\") " pod="openstack-operators/designate-operator-controller-manager-687df44cdb-9gn9l" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.596688 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4875b\" (UniqueName: \"kubernetes.io/projected/91f45294-2d2f-4b40-a1b4-a841c7748cc8-kube-api-access-4875b\") pod \"cinder-operator-controller-manager-59cdc64769-5glgp\" (UID: \"91f45294-2d2f-4b40-a1b4-a841c7748cc8\") " pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-5glgp" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.596736 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9kxq9\" (UniqueName: \"kubernetes.io/projected/6f3fbb7d-1207-45a2-8213-c7c71bd703f1-kube-api-access-9kxq9\") pod \"barbican-operator-controller-manager-64f84fcdbb-xz92m\" (UID: \"6f3fbb7d-1207-45a2-8213-c7c71bd703f1\") " pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-xz92m" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.602619 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d74794d9b-8ccrl"] Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.603489 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-8ccrl" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.609598 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"horizon-operator-controller-manager-dockercfg-8pn7w" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.610130 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d74794d9b-8ccrl"] Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.632835 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/infra-operator-controller-manager-685c45897-5gbwb"] Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.633860 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-685c45897-5gbwb" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.640093 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-webhook-server-cert" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.640414 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"infra-operator-controller-manager-dockercfg-mx224" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.640986 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-685c45897-5gbwb"] Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.644781 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ironic-operator-controller-manager-67494cb947-sklz6"] Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.645965 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-67494cb947-sklz6" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.649120 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-67494cb947-sklz6"] Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.649636 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ironic-operator-controller-manager-dockercfg-64q9z" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.697288 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/keystone-operator-controller-manager-ddb98f99b-brbrs"] Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.698435 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z7g9q\" (UniqueName: \"kubernetes.io/projected/2a2805ea-7d92-42e4-89cb-9b2672f72fc6-kube-api-access-z7g9q\") pod \"designate-operator-controller-manager-687df44cdb-9gn9l\" (UID: \"2a2805ea-7d92-42e4-89cb-9b2672f72fc6\") " pod="openstack-operators/designate-operator-controller-manager-687df44cdb-9gn9l" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.698480 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bcq5z\" (UniqueName: \"kubernetes.io/projected/ecd905f3-adb2-47ae-a86d-6fc0b4907a76-kube-api-access-bcq5z\") pod \"heat-operator-controller-manager-6d9967f8dd-ww7nt\" (UID: \"ecd905f3-adb2-47ae-a86d-6fc0b4907a76\") " pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-ww7nt" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.698514 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4875b\" (UniqueName: \"kubernetes.io/projected/91f45294-2d2f-4b40-a1b4-a841c7748cc8-kube-api-access-4875b\") pod \"cinder-operator-controller-manager-59cdc64769-5glgp\" (UID: \"91f45294-2d2f-4b40-a1b4-a841c7748cc8\") " pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-5glgp" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.698561 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5flr\" (UniqueName: \"kubernetes.io/projected/44eb39fa-5b20-4c02-b039-f23d7c5220d2-kube-api-access-h5flr\") pod \"glance-operator-controller-manager-7bb46cd7d-zt44b\" (UID: \"44eb39fa-5b20-4c02-b039-f23d7c5220d2\") " pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-zt44b" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.698603 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9kxq9\" (UniqueName: \"kubernetes.io/projected/6f3fbb7d-1207-45a2-8213-c7c71bd703f1-kube-api-access-9kxq9\") pod \"barbican-operator-controller-manager-64f84fcdbb-xz92m\" (UID: \"6f3fbb7d-1207-45a2-8213-c7c71bd703f1\") " pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-xz92m" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.702367 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-brbrs" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.704821 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"keystone-operator-controller-manager-dockercfg-b4jfr" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.736101 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z7g9q\" (UniqueName: \"kubernetes.io/projected/2a2805ea-7d92-42e4-89cb-9b2672f72fc6-kube-api-access-z7g9q\") pod \"designate-operator-controller-manager-687df44cdb-9gn9l\" (UID: \"2a2805ea-7d92-42e4-89cb-9b2672f72fc6\") " pod="openstack-operators/designate-operator-controller-manager-687df44cdb-9gn9l" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.744204 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/manila-operator-controller-manager-59578bc799-ls9k6"] Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.746253 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-59578bc799-ls9k6" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.752711 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9kxq9\" (UniqueName: \"kubernetes.io/projected/6f3fbb7d-1207-45a2-8213-c7c71bd703f1-kube-api-access-9kxq9\") pod \"barbican-operator-controller-manager-64f84fcdbb-xz92m\" (UID: \"6f3fbb7d-1207-45a2-8213-c7c71bd703f1\") " pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-xz92m" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.767007 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-xz92m" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.790987 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"manila-operator-controller-manager-dockercfg-kz4rd" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.800623 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5flr\" (UniqueName: \"kubernetes.io/projected/44eb39fa-5b20-4c02-b039-f23d7c5220d2-kube-api-access-h5flr\") pod \"glance-operator-controller-manager-7bb46cd7d-zt44b\" (UID: \"44eb39fa-5b20-4c02-b039-f23d7c5220d2\") " pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-zt44b" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.800712 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwqmt\" (UniqueName: \"kubernetes.io/projected/674852f1-9729-4a6e-8c90-31741fa69b49-kube-api-access-hwqmt\") pod \"ironic-operator-controller-manager-67494cb947-sklz6\" (UID: \"674852f1-9729-4a6e-8c90-31741fa69b49\") " pod="openstack-operators/ironic-operator-controller-manager-67494cb947-sklz6" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.800737 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6628a0ae-2a55-48b0-a5f1-0439c7d0c273-cert\") pod \"infra-operator-controller-manager-685c45897-5gbwb\" (UID: \"6628a0ae-2a55-48b0-a5f1-0439c7d0c273\") " pod="openstack-operators/infra-operator-controller-manager-685c45897-5gbwb" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.800760 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bcq5z\" (UniqueName: \"kubernetes.io/projected/ecd905f3-adb2-47ae-a86d-6fc0b4907a76-kube-api-access-bcq5z\") pod \"heat-operator-controller-manager-6d9967f8dd-ww7nt\" (UID: \"ecd905f3-adb2-47ae-a86d-6fc0b4907a76\") " pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-ww7nt" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.800784 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r2cl6\" (UniqueName: \"kubernetes.io/projected/e9feb54d-dbfa-40ec-ac45-535cca431024-kube-api-access-r2cl6\") pod \"horizon-operator-controller-manager-6d74794d9b-8ccrl\" (UID: \"e9feb54d-dbfa-40ec-ac45-535cca431024\") " pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-8ccrl" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.800809 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dbxkn\" (UniqueName: \"kubernetes.io/projected/950d3635-f160-45a1-9a4c-749720e662dc-kube-api-access-dbxkn\") pod \"keystone-operator-controller-manager-ddb98f99b-brbrs\" (UID: \"950d3635-f160-45a1-9a4c-749720e662dc\") " pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-brbrs" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.800833 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wchwh\" (UniqueName: \"kubernetes.io/projected/6628a0ae-2a55-48b0-a5f1-0439c7d0c273-kube-api-access-wchwh\") pod \"infra-operator-controller-manager-685c45897-5gbwb\" (UID: \"6628a0ae-2a55-48b0-a5f1-0439c7d0c273\") " pod="openstack-operators/infra-operator-controller-manager-685c45897-5gbwb" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.804031 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4875b\" (UniqueName: \"kubernetes.io/projected/91f45294-2d2f-4b40-a1b4-a841c7748cc8-kube-api-access-4875b\") pod \"cinder-operator-controller-manager-59cdc64769-5glgp\" (UID: \"91f45294-2d2f-4b40-a1b4-a841c7748cc8\") " pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-5glgp" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.804159 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-9gn9l" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.825346 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-ddb98f99b-brbrs"] Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.848264 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-59578bc799-ls9k6"] Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.848317 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5777b4f897-fjr8k"] Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.849490 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-fjr8k" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.856792 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5flr\" (UniqueName: \"kubernetes.io/projected/44eb39fa-5b20-4c02-b039-f23d7c5220d2-kube-api-access-h5flr\") pod \"glance-operator-controller-manager-7bb46cd7d-zt44b\" (UID: \"44eb39fa-5b20-4c02-b039-f23d7c5220d2\") " pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-zt44b" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.857174 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"mariadb-operator-controller-manager-dockercfg-d7kpb" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.858361 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bcq5z\" (UniqueName: \"kubernetes.io/projected/ecd905f3-adb2-47ae-a86d-6fc0b4907a76-kube-api-access-bcq5z\") pod \"heat-operator-controller-manager-6d9967f8dd-ww7nt\" (UID: \"ecd905f3-adb2-47ae-a86d-6fc0b4907a76\") " pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-ww7nt" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.864578 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/neutron-operator-controller-manager-797d478b46-6zmv5"] Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.865648 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-6zmv5" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.880562 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"neutron-operator-controller-manager-dockercfg-chxmt" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.881003 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5777b4f897-fjr8k"] Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.897097 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/nova-operator-controller-manager-57bb74c7bf-59px8"] Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.898123 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-59px8" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.905511 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"nova-operator-controller-manager-dockercfg-qvpvs" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.906024 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-797d478b46-6zmv5"] Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.907167 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwqmt\" (UniqueName: \"kubernetes.io/projected/674852f1-9729-4a6e-8c90-31741fa69b49-kube-api-access-hwqmt\") pod \"ironic-operator-controller-manager-67494cb947-sklz6\" (UID: \"674852f1-9729-4a6e-8c90-31741fa69b49\") " pod="openstack-operators/ironic-operator-controller-manager-67494cb947-sklz6" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.907194 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6628a0ae-2a55-48b0-a5f1-0439c7d0c273-cert\") pod \"infra-operator-controller-manager-685c45897-5gbwb\" (UID: \"6628a0ae-2a55-48b0-a5f1-0439c7d0c273\") " pod="openstack-operators/infra-operator-controller-manager-685c45897-5gbwb" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.907229 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r2cl6\" (UniqueName: \"kubernetes.io/projected/e9feb54d-dbfa-40ec-ac45-535cca431024-kube-api-access-r2cl6\") pod \"horizon-operator-controller-manager-6d74794d9b-8ccrl\" (UID: \"e9feb54d-dbfa-40ec-ac45-535cca431024\") " pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-8ccrl" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.907253 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dbxkn\" (UniqueName: \"kubernetes.io/projected/950d3635-f160-45a1-9a4c-749720e662dc-kube-api-access-dbxkn\") pod \"keystone-operator-controller-manager-ddb98f99b-brbrs\" (UID: \"950d3635-f160-45a1-9a4c-749720e662dc\") " pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-brbrs" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.907278 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wchwh\" (UniqueName: \"kubernetes.io/projected/6628a0ae-2a55-48b0-a5f1-0439c7d0c273-kube-api-access-wchwh\") pod \"infra-operator-controller-manager-685c45897-5gbwb\" (UID: \"6628a0ae-2a55-48b0-a5f1-0439c7d0c273\") " pod="openstack-operators/infra-operator-controller-manager-685c45897-5gbwb" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.907298 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bn89b\" (UniqueName: \"kubernetes.io/projected/20584c10-d8a1-41ca-a0a2-9465fa5190f2-kube-api-access-bn89b\") pod \"manila-operator-controller-manager-59578bc799-ls9k6\" (UID: \"20584c10-d8a1-41ca-a0a2-9465fa5190f2\") " pod="openstack-operators/manila-operator-controller-manager-59578bc799-ls9k6" Oct 11 04:04:47 crc kubenswrapper[4967]: E1011 04:04:47.908629 4967 secret.go:188] Couldn't get secret openstack-operators/infra-operator-webhook-server-cert: secret "infra-operator-webhook-server-cert" not found Oct 11 04:04:47 crc kubenswrapper[4967]: E1011 04:04:47.908686 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/6628a0ae-2a55-48b0-a5f1-0439c7d0c273-cert podName:6628a0ae-2a55-48b0-a5f1-0439c7d0c273 nodeName:}" failed. No retries permitted until 2025-10-11 04:04:48.408670865 +0000 UTC m=+816.371879798 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/6628a0ae-2a55-48b0-a5f1-0439c7d0c273-cert") pod "infra-operator-controller-manager-685c45897-5gbwb" (UID: "6628a0ae-2a55-48b0-a5f1-0439c7d0c273") : secret "infra-operator-webhook-server-cert" not found Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.909110 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-ww7nt" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.929286 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwqmt\" (UniqueName: \"kubernetes.io/projected/674852f1-9729-4a6e-8c90-31741fa69b49-kube-api-access-hwqmt\") pod \"ironic-operator-controller-manager-67494cb947-sklz6\" (UID: \"674852f1-9729-4a6e-8c90-31741fa69b49\") " pod="openstack-operators/ironic-operator-controller-manager-67494cb947-sklz6" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.932675 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-7clg2"] Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.933930 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-7clg2" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.935599 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wchwh\" (UniqueName: \"kubernetes.io/projected/6628a0ae-2a55-48b0-a5f1-0439c7d0c273-kube-api-access-wchwh\") pod \"infra-operator-controller-manager-685c45897-5gbwb\" (UID: \"6628a0ae-2a55-48b0-a5f1-0439c7d0c273\") " pod="openstack-operators/infra-operator-controller-manager-685c45897-5gbwb" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.944999 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"octavia-operator-controller-manager-dockercfg-6t6fn" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.950943 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-57bb74c7bf-59px8"] Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.954617 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r2cl6\" (UniqueName: \"kubernetes.io/projected/e9feb54d-dbfa-40ec-ac45-535cca431024-kube-api-access-r2cl6\") pod \"horizon-operator-controller-manager-6d74794d9b-8ccrl\" (UID: \"e9feb54d-dbfa-40ec-ac45-535cca431024\") " pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-8ccrl" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.967752 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dbxkn\" (UniqueName: \"kubernetes.io/projected/950d3635-f160-45a1-9a4c-749720e662dc-kube-api-access-dbxkn\") pod \"keystone-operator-controller-manager-ddb98f99b-brbrs\" (UID: \"950d3635-f160-45a1-9a4c-749720e662dc\") " pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-brbrs" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.971917 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-7clg2"] Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.985198 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns"] Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.986329 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.987441 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ironic-operator-controller-manager-67494cb947-sklz6" Oct 11 04:04:47 crc kubenswrapper[4967]: I1011 04:04:47.998438 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-webhook-server-cert" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.009158 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-baremetal-operator-controller-manager-dockercfg-pdvcp" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.010695 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ldjzg\" (UniqueName: \"kubernetes.io/projected/7140d0e1-9937-491b-aa8a-16db830aff11-kube-api-access-ldjzg\") pod \"nova-operator-controller-manager-57bb74c7bf-59px8\" (UID: \"7140d0e1-9937-491b-aa8a-16db830aff11\") " pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-59px8" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.010768 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bn89b\" (UniqueName: \"kubernetes.io/projected/20584c10-d8a1-41ca-a0a2-9465fa5190f2-kube-api-access-bn89b\") pod \"manila-operator-controller-manager-59578bc799-ls9k6\" (UID: \"20584c10-d8a1-41ca-a0a2-9465fa5190f2\") " pod="openstack-operators/manila-operator-controller-manager-59578bc799-ls9k6" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.010823 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2psf9\" (UniqueName: \"kubernetes.io/projected/538ebbcb-d51c-472c-a2b4-80236b0c8c08-kube-api-access-2psf9\") pod \"neutron-operator-controller-manager-797d478b46-6zmv5\" (UID: \"538ebbcb-d51c-472c-a2b4-80236b0c8c08\") " pod="openstack-operators/neutron-operator-controller-manager-797d478b46-6zmv5" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.010841 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tgnl4\" (UniqueName: \"kubernetes.io/projected/ef8ae8f4-ad95-4f1f-bc8c-170fe5058dc9-kube-api-access-tgnl4\") pod \"mariadb-operator-controller-manager-5777b4f897-fjr8k\" (UID: \"ef8ae8f4-ad95-4f1f-bc8c-170fe5058dc9\") " pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-fjr8k" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.018607 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/ovn-operator-controller-manager-869cc7797f-kkrkc"] Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.020064 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-kkrkc" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.040141 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"ovn-operator-controller-manager-dockercfg-gw6bm" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.040321 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/placement-operator-controller-manager-664664cb68-vng4c"] Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.060125 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-664664cb68-vng4c" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.077444 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-5glgp" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.077914 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-869cc7797f-kkrkc"] Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.078444 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bn89b\" (UniqueName: \"kubernetes.io/projected/20584c10-d8a1-41ca-a0a2-9465fa5190f2-kube-api-access-bn89b\") pod \"manila-operator-controller-manager-59578bc799-ls9k6\" (UID: \"20584c10-d8a1-41ca-a0a2-9465fa5190f2\") " pod="openstack-operators/manila-operator-controller-manager-59578bc799-ls9k6" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.078751 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"placement-operator-controller-manager-dockercfg-4lknl" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.111697 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns"] Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.120915 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lw5cp\" (UniqueName: \"kubernetes.io/projected/26b9d5a0-dfbe-4e53-aa03-4fbb4682019b-kube-api-access-lw5cp\") pod \"placement-operator-controller-manager-664664cb68-vng4c\" (UID: \"26b9d5a0-dfbe-4e53-aa03-4fbb4682019b\") " pod="openstack-operators/placement-operator-controller-manager-664664cb68-vng4c" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.120955 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ldjzg\" (UniqueName: \"kubernetes.io/projected/7140d0e1-9937-491b-aa8a-16db830aff11-kube-api-access-ldjzg\") pod \"nova-operator-controller-manager-57bb74c7bf-59px8\" (UID: \"7140d0e1-9937-491b-aa8a-16db830aff11\") " pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-59px8" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.121007 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0cc11444-d341-4821-a743-6b4af1466e4a-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns\" (UID: \"0cc11444-d341-4821-a743-6b4af1466e4a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.121028 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2psf9\" (UniqueName: \"kubernetes.io/projected/538ebbcb-d51c-472c-a2b4-80236b0c8c08-kube-api-access-2psf9\") pod \"neutron-operator-controller-manager-797d478b46-6zmv5\" (UID: \"538ebbcb-d51c-472c-a2b4-80236b0c8c08\") " pod="openstack-operators/neutron-operator-controller-manager-797d478b46-6zmv5" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.121043 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tgnl4\" (UniqueName: \"kubernetes.io/projected/ef8ae8f4-ad95-4f1f-bc8c-170fe5058dc9-kube-api-access-tgnl4\") pod \"mariadb-operator-controller-manager-5777b4f897-fjr8k\" (UID: \"ef8ae8f4-ad95-4f1f-bc8c-170fe5058dc9\") " pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-fjr8k" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.121090 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hql9t\" (UniqueName: \"kubernetes.io/projected/0cc11444-d341-4821-a743-6b4af1466e4a-kube-api-access-hql9t\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns\" (UID: \"0cc11444-d341-4821-a743-6b4af1466e4a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.121127 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g9j5n\" (UniqueName: \"kubernetes.io/projected/7e917a6a-9720-4427-b079-33d112a981a3-kube-api-access-g9j5n\") pod \"octavia-operator-controller-manager-6d7c7ddf95-7clg2\" (UID: \"7e917a6a-9720-4427-b079-33d112a981a3\") " pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-7clg2" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.156353 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-zt44b" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.160721 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-brbrs" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.174738 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ldjzg\" (UniqueName: \"kubernetes.io/projected/7140d0e1-9937-491b-aa8a-16db830aff11-kube-api-access-ldjzg\") pod \"nova-operator-controller-manager-57bb74c7bf-59px8\" (UID: \"7140d0e1-9937-491b-aa8a-16db830aff11\") " pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-59px8" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.175242 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tgnl4\" (UniqueName: \"kubernetes.io/projected/ef8ae8f4-ad95-4f1f-bc8c-170fe5058dc9-kube-api-access-tgnl4\") pod \"mariadb-operator-controller-manager-5777b4f897-fjr8k\" (UID: \"ef8ae8f4-ad95-4f1f-bc8c-170fe5058dc9\") " pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-fjr8k" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.176419 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2psf9\" (UniqueName: \"kubernetes.io/projected/538ebbcb-d51c-472c-a2b4-80236b0c8c08-kube-api-access-2psf9\") pod \"neutron-operator-controller-manager-797d478b46-6zmv5\" (UID: \"538ebbcb-d51c-472c-a2b4-80236b0c8c08\") " pod="openstack-operators/neutron-operator-controller-manager-797d478b46-6zmv5" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.177778 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-664664cb68-vng4c"] Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.214995 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-65knj"] Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.216120 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-65knj" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.222173 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0cc11444-d341-4821-a743-6b4af1466e4a-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns\" (UID: \"0cc11444-d341-4821-a743-6b4af1466e4a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.222219 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dj8bk\" (UniqueName: \"kubernetes.io/projected/1094a8be-16bb-4986-b4a4-465d41880800-kube-api-access-dj8bk\") pod \"swift-operator-controller-manager-5f4d5dfdc6-65knj\" (UID: \"1094a8be-16bb-4986-b4a4-465d41880800\") " pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-65knj" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.222250 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hql9t\" (UniqueName: \"kubernetes.io/projected/0cc11444-d341-4821-a743-6b4af1466e4a-kube-api-access-hql9t\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns\" (UID: \"0cc11444-d341-4821-a743-6b4af1466e4a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.222288 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g9j5n\" (UniqueName: \"kubernetes.io/projected/7e917a6a-9720-4427-b079-33d112a981a3-kube-api-access-g9j5n\") pod \"octavia-operator-controller-manager-6d7c7ddf95-7clg2\" (UID: \"7e917a6a-9720-4427-b079-33d112a981a3\") " pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-7clg2" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.222313 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lw5cp\" (UniqueName: \"kubernetes.io/projected/26b9d5a0-dfbe-4e53-aa03-4fbb4682019b-kube-api-access-lw5cp\") pod \"placement-operator-controller-manager-664664cb68-vng4c\" (UID: \"26b9d5a0-dfbe-4e53-aa03-4fbb4682019b\") " pod="openstack-operators/placement-operator-controller-manager-664664cb68-vng4c" Oct 11 04:04:48 crc kubenswrapper[4967]: E1011 04:04:48.222338 4967 secret.go:188] Couldn't get secret openstack-operators/openstack-baremetal-operator-webhook-server-cert: secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.222359 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xmds\" (UniqueName: \"kubernetes.io/projected/c9e29e65-cdd1-408d-bed0-bc283147b283-kube-api-access-4xmds\") pod \"ovn-operator-controller-manager-869cc7797f-kkrkc\" (UID: \"c9e29e65-cdd1-408d-bed0-bc283147b283\") " pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-kkrkc" Oct 11 04:04:48 crc kubenswrapper[4967]: E1011 04:04:48.222399 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/0cc11444-d341-4821-a743-6b4af1466e4a-cert podName:0cc11444-d341-4821-a743-6b4af1466e4a nodeName:}" failed. No retries permitted until 2025-10-11 04:04:48.722382142 +0000 UTC m=+816.685591075 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "cert" (UniqueName: "kubernetes.io/secret/0cc11444-d341-4821-a743-6b4af1466e4a-cert") pod "openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns" (UID: "0cc11444-d341-4821-a743-6b4af1466e4a") : secret "openstack-baremetal-operator-webhook-server-cert" not found Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.226129 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-8ccrl" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.226474 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/manila-operator-controller-manager-59578bc799-ls9k6" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.238277 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-65knj"] Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.238548 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"swift-operator-controller-manager-dockercfg-4s7p6" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.248761 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lw5cp\" (UniqueName: \"kubernetes.io/projected/26b9d5a0-dfbe-4e53-aa03-4fbb4682019b-kube-api-access-lw5cp\") pod \"placement-operator-controller-manager-664664cb68-vng4c\" (UID: \"26b9d5a0-dfbe-4e53-aa03-4fbb4682019b\") " pod="openstack-operators/placement-operator-controller-manager-664664cb68-vng4c" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.249266 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-fjr8k" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.256609 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hql9t\" (UniqueName: \"kubernetes.io/projected/0cc11444-d341-4821-a743-6b4af1466e4a-kube-api-access-hql9t\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns\" (UID: \"0cc11444-d341-4821-a743-6b4af1466e4a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.291083 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g9j5n\" (UniqueName: \"kubernetes.io/projected/7e917a6a-9720-4427-b079-33d112a981a3-kube-api-access-g9j5n\") pod \"octavia-operator-controller-manager-6d7c7ddf95-7clg2\" (UID: \"7e917a6a-9720-4427-b079-33d112a981a3\") " pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-7clg2" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.301667 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-6zmv5" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.316956 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-578874c84d-fgz82"] Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.318006 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-fgz82" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.323248 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dj8bk\" (UniqueName: \"kubernetes.io/projected/1094a8be-16bb-4986-b4a4-465d41880800-kube-api-access-dj8bk\") pod \"swift-operator-controller-manager-5f4d5dfdc6-65knj\" (UID: \"1094a8be-16bb-4986-b4a4-465d41880800\") " pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-65knj" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.323444 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xmds\" (UniqueName: \"kubernetes.io/projected/c9e29e65-cdd1-408d-bed0-bc283147b283-kube-api-access-4xmds\") pod \"ovn-operator-controller-manager-869cc7797f-kkrkc\" (UID: \"c9e29e65-cdd1-408d-bed0-bc283147b283\") " pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-kkrkc" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.323732 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-59px8" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.330402 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"telemetry-operator-controller-manager-dockercfg-wj84w" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.352654 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dj8bk\" (UniqueName: \"kubernetes.io/projected/1094a8be-16bb-4986-b4a4-465d41880800-kube-api-access-dj8bk\") pod \"swift-operator-controller-manager-5f4d5dfdc6-65knj\" (UID: \"1094a8be-16bb-4986-b4a4-465d41880800\") " pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-65knj" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.352720 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-578874c84d-fgz82"] Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.353246 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xmds\" (UniqueName: \"kubernetes.io/projected/c9e29e65-cdd1-408d-bed0-bc283147b283-kube-api-access-4xmds\") pod \"ovn-operator-controller-manager-869cc7797f-kkrkc\" (UID: \"c9e29e65-cdd1-408d-bed0-bc283147b283\") " pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-kkrkc" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.377320 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/test-operator-controller-manager-ffcdd6c94-4s7hv"] Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.379016 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-4s7hv" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.379959 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-7clg2" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.384005 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"test-operator-controller-manager-dockercfg-b7s6n" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.410548 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-9gn9l" event={"ID":"2a2805ea-7d92-42e4-89cb-9b2672f72fc6","Type":"ContainerStarted","Data":"08fbefcec471ac0571d6eb516d4649fc05d85302e14e3586499d1c6caf5c0b4f"} Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.433817 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-kkrkc" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.438228 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f76h4\" (UniqueName: \"kubernetes.io/projected/5e24d70e-8882-45c0-9f6b-e443549ecabb-kube-api-access-f76h4\") pod \"test-operator-controller-manager-ffcdd6c94-4s7hv\" (UID: \"5e24d70e-8882-45c0-9f6b-e443549ecabb\") " pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-4s7hv" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.438342 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2dp7h\" (UniqueName: \"kubernetes.io/projected/e82102ad-391b-4b29-ae47-dae18dba7743-kube-api-access-2dp7h\") pod \"telemetry-operator-controller-manager-578874c84d-fgz82\" (UID: \"e82102ad-391b-4b29-ae47-dae18dba7743\") " pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-fgz82" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.438383 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6628a0ae-2a55-48b0-a5f1-0439c7d0c273-cert\") pod \"infra-operator-controller-manager-685c45897-5gbwb\" (UID: \"6628a0ae-2a55-48b0-a5f1-0439c7d0c273\") " pod="openstack-operators/infra-operator-controller-manager-685c45897-5gbwb" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.441910 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-ffcdd6c94-4s7hv"] Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.490325 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/placement-operator-controller-manager-664664cb68-vng4c" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.497668 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/6628a0ae-2a55-48b0-a5f1-0439c7d0c273-cert\") pod \"infra-operator-controller-manager-685c45897-5gbwb\" (UID: \"6628a0ae-2a55-48b0-a5f1-0439c7d0c273\") " pod="openstack-operators/infra-operator-controller-manager-685c45897-5gbwb" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.511918 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/watcher-operator-controller-manager-646675d848-zdpbn"] Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.522229 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-646675d848-zdpbn" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.528863 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"watcher-operator-controller-manager-dockercfg-2hd7d" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.538016 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-646675d848-zdpbn"] Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.539734 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lxkjt\" (UniqueName: \"kubernetes.io/projected/10a0ea36-9248-4086-83e4-de8578492421-kube-api-access-lxkjt\") pod \"watcher-operator-controller-manager-646675d848-zdpbn\" (UID: \"10a0ea36-9248-4086-83e4-de8578492421\") " pod="openstack-operators/watcher-operator-controller-manager-646675d848-zdpbn" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.539816 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f76h4\" (UniqueName: \"kubernetes.io/projected/5e24d70e-8882-45c0-9f6b-e443549ecabb-kube-api-access-f76h4\") pod \"test-operator-controller-manager-ffcdd6c94-4s7hv\" (UID: \"5e24d70e-8882-45c0-9f6b-e443549ecabb\") " pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-4s7hv" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.540711 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2dp7h\" (UniqueName: \"kubernetes.io/projected/e82102ad-391b-4b29-ae47-dae18dba7743-kube-api-access-2dp7h\") pod \"telemetry-operator-controller-manager-578874c84d-fgz82\" (UID: \"e82102ad-391b-4b29-ae47-dae18dba7743\") " pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-fgz82" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.556024 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/infra-operator-controller-manager-685c45897-5gbwb" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.559984 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f76h4\" (UniqueName: \"kubernetes.io/projected/5e24d70e-8882-45c0-9f6b-e443549ecabb-kube-api-access-f76h4\") pod \"test-operator-controller-manager-ffcdd6c94-4s7hv\" (UID: \"5e24d70e-8882-45c0-9f6b-e443549ecabb\") " pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-4s7hv" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.562185 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2dp7h\" (UniqueName: \"kubernetes.io/projected/e82102ad-391b-4b29-ae47-dae18dba7743-kube-api-access-2dp7h\") pod \"telemetry-operator-controller-manager-578874c84d-fgz82\" (UID: \"e82102ad-391b-4b29-ae47-dae18dba7743\") " pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-fgz82" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.590856 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/openstack-operator-controller-manager-66664b49c9-tk2lg"] Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.592394 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-66664b49c9-tk2lg" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.596389 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-65knj" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.600011 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"openstack-operator-controller-manager-dockercfg-wtzql" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.600215 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"webhook-server-cert" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.607355 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/designate-operator-controller-manager-687df44cdb-9gn9l"] Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.624716 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-66664b49c9-tk2lg"] Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.638161 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-cmzk2"] Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.639046 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-cmzk2" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.642733 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lxkjt\" (UniqueName: \"kubernetes.io/projected/10a0ea36-9248-4086-83e4-de8578492421-kube-api-access-lxkjt\") pod \"watcher-operator-controller-manager-646675d848-zdpbn\" (UID: \"10a0ea36-9248-4086-83e4-de8578492421\") " pod="openstack-operators/watcher-operator-controller-manager-646675d848-zdpbn" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.644475 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack-operators"/"rabbitmq-cluster-operator-controller-manager-dockercfg-6kk8n" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.644626 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-cmzk2"] Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.674112 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lxkjt\" (UniqueName: \"kubernetes.io/projected/10a0ea36-9248-4086-83e4-de8578492421-kube-api-access-lxkjt\") pod \"watcher-operator-controller-manager-646675d848-zdpbn\" (UID: \"10a0ea36-9248-4086-83e4-de8578492421\") " pod="openstack-operators/watcher-operator-controller-manager-646675d848-zdpbn" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.743880 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cpwn8\" (UniqueName: \"kubernetes.io/projected/11e4f97c-6107-45fd-8eb1-e39ef7bed4ce-kube-api-access-cpwn8\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-cmzk2\" (UID: \"11e4f97c-6107-45fd-8eb1-e39ef7bed4ce\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-cmzk2" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.743940 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xv8gj\" (UniqueName: \"kubernetes.io/projected/620877eb-6f9c-4647-b545-1cca7a438690-kube-api-access-xv8gj\") pod \"openstack-operator-controller-manager-66664b49c9-tk2lg\" (UID: \"620877eb-6f9c-4647-b545-1cca7a438690\") " pod="openstack-operators/openstack-operator-controller-manager-66664b49c9-tk2lg" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.743973 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0cc11444-d341-4821-a743-6b4af1466e4a-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns\" (UID: \"0cc11444-d341-4821-a743-6b4af1466e4a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.744001 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/620877eb-6f9c-4647-b545-1cca7a438690-cert\") pod \"openstack-operator-controller-manager-66664b49c9-tk2lg\" (UID: \"620877eb-6f9c-4647-b545-1cca7a438690\") " pod="openstack-operators/openstack-operator-controller-manager-66664b49c9-tk2lg" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.746843 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-fgz82" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.754427 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/0cc11444-d341-4821-a743-6b4af1466e4a-cert\") pod \"openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns\" (UID: \"0cc11444-d341-4821-a743-6b4af1466e4a\") " pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.759488 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/barbican-operator-controller-manager-64f84fcdbb-xz92m"] Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.796405 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-4s7hv" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.847380 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cpwn8\" (UniqueName: \"kubernetes.io/projected/11e4f97c-6107-45fd-8eb1-e39ef7bed4ce-kube-api-access-cpwn8\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-cmzk2\" (UID: \"11e4f97c-6107-45fd-8eb1-e39ef7bed4ce\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-cmzk2" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.847638 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xv8gj\" (UniqueName: \"kubernetes.io/projected/620877eb-6f9c-4647-b545-1cca7a438690-kube-api-access-xv8gj\") pod \"openstack-operator-controller-manager-66664b49c9-tk2lg\" (UID: \"620877eb-6f9c-4647-b545-1cca7a438690\") " pod="openstack-operators/openstack-operator-controller-manager-66664b49c9-tk2lg" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.847685 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/620877eb-6f9c-4647-b545-1cca7a438690-cert\") pod \"openstack-operator-controller-manager-66664b49c9-tk2lg\" (UID: \"620877eb-6f9c-4647-b545-1cca7a438690\") " pod="openstack-operators/openstack-operator-controller-manager-66664b49c9-tk2lg" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.859084 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/620877eb-6f9c-4647-b545-1cca7a438690-cert\") pod \"openstack-operator-controller-manager-66664b49c9-tk2lg\" (UID: \"620877eb-6f9c-4647-b545-1cca7a438690\") " pod="openstack-operators/openstack-operator-controller-manager-66664b49c9-tk2lg" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.862528 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/watcher-operator-controller-manager-646675d848-zdpbn" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.872100 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cpwn8\" (UniqueName: \"kubernetes.io/projected/11e4f97c-6107-45fd-8eb1-e39ef7bed4ce-kube-api-access-cpwn8\") pod \"rabbitmq-cluster-operator-manager-5f97d8c699-cmzk2\" (UID: \"11e4f97c-6107-45fd-8eb1-e39ef7bed4ce\") " pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-cmzk2" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.877469 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xv8gj\" (UniqueName: \"kubernetes.io/projected/620877eb-6f9c-4647-b545-1cca7a438690-kube-api-access-xv8gj\") pod \"openstack-operator-controller-manager-66664b49c9-tk2lg\" (UID: \"620877eb-6f9c-4647-b545-1cca7a438690\") " pod="openstack-operators/openstack-operator-controller-manager-66664b49c9-tk2lg" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.946564 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-operator-controller-manager-66664b49c9-tk2lg" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.964115 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/glance-operator-controller-manager-7bb46cd7d-zt44b"] Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.964456 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-cmzk2" Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.970126 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/cinder-operator-controller-manager-59cdc64769-5glgp"] Oct 11 04:04:48 crc kubenswrapper[4967]: I1011 04:04:48.977115 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ironic-operator-controller-manager-67494cb947-sklz6"] Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.004541 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/heat-operator-controller-manager-6d9967f8dd-ww7nt"] Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.013933 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns" Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.064776 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-8nwxv" Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.065038 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-8nwxv" Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.120822 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-8nwxv" Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.174359 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/horizon-operator-controller-manager-6d74794d9b-8ccrl"] Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.179204 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/manila-operator-controller-manager-59578bc799-ls9k6"] Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.206418 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/mariadb-operator-controller-manager-5777b4f897-fjr8k"] Oct 11 04:04:49 crc kubenswrapper[4967]: W1011 04:04:49.216450 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef8ae8f4_ad95_4f1f_bc8c_170fe5058dc9.slice/crio-1e64ad31ce00162f8f86beac77f45f8d82e47fbf27157f4b4f22f429646b505a WatchSource:0}: Error finding container 1e64ad31ce00162f8f86beac77f45f8d82e47fbf27157f4b4f22f429646b505a: Status 404 returned error can't find the container with id 1e64ad31ce00162f8f86beac77f45f8d82e47fbf27157f4b4f22f429646b505a Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.333939 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/keystone-operator-controller-manager-ddb98f99b-brbrs"] Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.339834 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-7clg2"] Oct 11 04:04:49 crc kubenswrapper[4967]: W1011 04:04:49.340588 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod950d3635_f160_45a1_9a4c_749720e662dc.slice/crio-0b181cbf19495808e6fd6362cd1c66fc223a445ca8a606976667ef26a3694199 WatchSource:0}: Error finding container 0b181cbf19495808e6fd6362cd1c66fc223a445ca8a606976667ef26a3694199: Status 404 returned error can't find the container with id 0b181cbf19495808e6fd6362cd1c66fc223a445ca8a606976667ef26a3694199 Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.345130 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/ovn-operator-controller-manager-869cc7797f-kkrkc"] Oct 11 04:04:49 crc kubenswrapper[4967]: W1011 04:04:49.346048 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc9e29e65_cdd1_408d_bed0_bc283147b283.slice/crio-451b96c28bcc3cef32d8632abe9bef49ff5fa3f8c25ab68bd1d9dc73a751b5a7 WatchSource:0}: Error finding container 451b96c28bcc3cef32d8632abe9bef49ff5fa3f8c25ab68bd1d9dc73a751b5a7: Status 404 returned error can't find the container with id 451b96c28bcc3cef32d8632abe9bef49ff5fa3f8c25ab68bd1d9dc73a751b5a7 Oct 11 04:04:49 crc kubenswrapper[4967]: W1011 04:04:49.355984 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e917a6a_9720_4427_b079_33d112a981a3.slice/crio-8ce806dd2704e31027b7fb2a6976c38dc63ffb118cfb8a069bac21a890e45a40 WatchSource:0}: Error finding container 8ce806dd2704e31027b7fb2a6976c38dc63ffb118cfb8a069bac21a890e45a40: Status 404 returned error can't find the container with id 8ce806dd2704e31027b7fb2a6976c38dc63ffb118cfb8a069bac21a890e45a40 Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.359846 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/nova-operator-controller-manager-57bb74c7bf-59px8"] Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.420230 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-59px8" event={"ID":"7140d0e1-9937-491b-aa8a-16db830aff11","Type":"ContainerStarted","Data":"743779140c58f36feadc6cf7ff5384db7b9048b2c5aa1b84ce748db6b5c19b48"} Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.422702 4967 generic.go:334] "Generic (PLEG): container finished" podID="7de04fa6-4dd6-4254-bde0-b83b920be0fa" containerID="f6a5c5ba5a22010cfe674fced11a2d28812f088ff6b617ffd64a731a3a04616a" exitCode=0 Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.422751 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z9scd" event={"ID":"7de04fa6-4dd6-4254-bde0-b83b920be0fa","Type":"ContainerDied","Data":"f6a5c5ba5a22010cfe674fced11a2d28812f088ff6b617ffd64a731a3a04616a"} Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.426629 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67494cb947-sklz6" event={"ID":"674852f1-9729-4a6e-8c90-31741fa69b49","Type":"ContainerStarted","Data":"651e1e1f42fb7e3eec5c62bc6aac916c130ae60bf02d16d0a55ce1c981635771"} Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.428557 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-zt44b" event={"ID":"44eb39fa-5b20-4c02-b039-f23d7c5220d2","Type":"ContainerStarted","Data":"75e42402cdfaee7b39c56d61f90b9272acb75d670ba8114ec5627aa0ca930cbc"} Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.432094 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-7clg2" event={"ID":"7e917a6a-9720-4427-b079-33d112a981a3","Type":"ContainerStarted","Data":"8ce806dd2704e31027b7fb2a6976c38dc63ffb118cfb8a069bac21a890e45a40"} Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.433261 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-kkrkc" event={"ID":"c9e29e65-cdd1-408d-bed0-bc283147b283","Type":"ContainerStarted","Data":"451b96c28bcc3cef32d8632abe9bef49ff5fa3f8c25ab68bd1d9dc73a751b5a7"} Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.435211 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-brbrs" event={"ID":"950d3635-f160-45a1-9a4c-749720e662dc","Type":"ContainerStarted","Data":"0b181cbf19495808e6fd6362cd1c66fc223a445ca8a606976667ef26a3694199"} Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.436113 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-fjr8k" event={"ID":"ef8ae8f4-ad95-4f1f-bc8c-170fe5058dc9","Type":"ContainerStarted","Data":"1e64ad31ce00162f8f86beac77f45f8d82e47fbf27157f4b4f22f429646b505a"} Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.440134 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-8ccrl" event={"ID":"e9feb54d-dbfa-40ec-ac45-535cca431024","Type":"ContainerStarted","Data":"46104522bf7066d6ff1c8125fd9df477873d1f60513273a4856e7c6ca9c3b10c"} Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.441231 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-ww7nt" event={"ID":"ecd905f3-adb2-47ae-a86d-6fc0b4907a76","Type":"ContainerStarted","Data":"c9e6a3b82428ee627f16704cbbbc78f81d9616b0c63bb6f51bfa4ce26f11f0af"} Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.442147 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-xz92m" event={"ID":"6f3fbb7d-1207-45a2-8213-c7c71bd703f1","Type":"ContainerStarted","Data":"2367a1737af426d91c33d9cb6b33958a5d37bf33736fe9c1398eff3e4c71911f"} Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.452716 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-59578bc799-ls9k6" event={"ID":"20584c10-d8a1-41ca-a0a2-9465fa5190f2","Type":"ContainerStarted","Data":"1dbf3ed079c793a501d0b7b039a15d9e73aefb13f67679d8c3d0f0ffb274d5e7"} Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.460484 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-5glgp" event={"ID":"91f45294-2d2f-4b40-a1b4-a841c7748cc8","Type":"ContainerStarted","Data":"8263a8e08b2da93903dbcf87befbc520de07b425eac83c655f8ff4f59b7a931d"} Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.521803 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-8nwxv" Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.732090 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/telemetry-operator-controller-manager-578874c84d-fgz82"] Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.750175 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/neutron-operator-controller-manager-797d478b46-6zmv5"] Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.765588 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/placement-operator-controller-manager-664664cb68-vng4c"] Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.772229 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/watcher-operator-controller-manager-646675d848-zdpbn"] Oct 11 04:04:49 crc kubenswrapper[4967]: E1011 04:04:49.775510 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/telemetry-operator@sha256:abe978f8da75223de5043cca50278ad4e28c8dd309883f502fe1e7a9998733b0,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-2dp7h,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod telemetry-operator-controller-manager-578874c84d-fgz82_openstack-operators(e82102ad-391b-4b29-ae47-dae18dba7743): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.778172 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/infra-operator-controller-manager-685c45897-5gbwb"] Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.789180 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/test-operator-controller-manager-ffcdd6c94-4s7hv"] Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.794236 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-cmzk2"] Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.799686 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-65knj"] Oct 11 04:04:49 crc kubenswrapper[4967]: E1011 04:04:49.803963 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lw5cp,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod placement-operator-controller-manager-664664cb68-vng4c_openstack-operators(26b9d5a0-dfbe-4e53-aa03-4fbb4682019b): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 11 04:04:49 crc kubenswrapper[4967]: W1011 04:04:49.806615 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod10a0ea36_9248_4086_83e4_de8578492421.slice/crio-fccba244efb965e21ea982f9f9058efcbd07e08bb723fa4f4fb64dc750a1957a WatchSource:0}: Error finding container fccba244efb965e21ea982f9f9058efcbd07e08bb723fa4f4fb64dc750a1957a: Status 404 returned error can't find the container with id fccba244efb965e21ea982f9f9058efcbd07e08bb723fa4f4fb64dc750a1957a Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.808410 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns"] Oct 11 04:04:49 crc kubenswrapper[4967]: E1011 04:04:49.810460 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/test-operator@sha256:7e584b1c430441c8b6591dadeff32e065de8a185ad37ef90d2e08d37e59aab4a,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-f76h4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod test-operator-controller-manager-ffcdd6c94-4s7hv_openstack-operators(5e24d70e-8882-45c0-9f6b-e443549ecabb): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 11 04:04:49 crc kubenswrapper[4967]: E1011 04:04:49.812862 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/watcher-operator@sha256:98a5233f0596591acdf2c6a5838b08be108787cdb6ad1995b2b7886bac0fe6ca,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-lxkjt,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod watcher-operator-controller-manager-646675d848-zdpbn_openstack-operators(10a0ea36-9248-4086-83e4-de8578492421): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 11 04:04:49 crc kubenswrapper[4967]: I1011 04:04:49.816007 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack-operators/openstack-operator-controller-manager-66664b49c9-tk2lg"] Oct 11 04:04:49 crc kubenswrapper[4967]: W1011 04:04:49.832149 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod1094a8be_16bb_4986_b4a4_465d41880800.slice/crio-182891a6f5158a0fbe8e9dad5d953309eae6da2178ea3d81d41e30d4d5ce6cf7 WatchSource:0}: Error finding container 182891a6f5158a0fbe8e9dad5d953309eae6da2178ea3d81d41e30d4d5ce6cf7: Status 404 returned error can't find the container with id 182891a6f5158a0fbe8e9dad5d953309eae6da2178ea3d81d41e30d4d5ce6cf7 Oct 11 04:04:49 crc kubenswrapper[4967]: W1011 04:04:49.837880 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod0cc11444_d341_4821_a743_6b4af1466e4a.slice/crio-9089b2f32b207344bccf3e939d8aafe58cf76491363672b10f2d39c7b4823e1f WatchSource:0}: Error finding container 9089b2f32b207344bccf3e939d8aafe58cf76491363672b10f2d39c7b4823e1f: Status 404 returned error can't find the container with id 9089b2f32b207344bccf3e939d8aafe58cf76491363672b10f2d39c7b4823e1f Oct 11 04:04:49 crc kubenswrapper[4967]: W1011 04:04:49.843656 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod620877eb_6f9c_4647_b545_1cca7a438690.slice/crio-9b30a12c127915cec43a5046c277114291467eb194afb2e31ba67ec2bf13e2a8 WatchSource:0}: Error finding container 9b30a12c127915cec43a5046c277114291467eb194afb2e31ba67ec2bf13e2a8: Status 404 returned error can't find the container with id 9b30a12c127915cec43a5046c277114291467eb194afb2e31ba67ec2bf13e2a8 Oct 11 04:04:49 crc kubenswrapper[4967]: E1011 04:04:49.843866 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dj8bk,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod swift-operator-controller-manager-5f4d5dfdc6-65knj_openstack-operators(1094a8be-16bb-4986-b4a4-465d41880800): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 11 04:04:49 crc kubenswrapper[4967]: W1011 04:04:49.844296 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod11e4f97c_6107_45fd_8eb1_e39ef7bed4ce.slice/crio-d4b21179521e1a0c0b5a6e3a71b59be1c7be006f0cfaaf508bde6ac07c811391 WatchSource:0}: Error finding container d4b21179521e1a0c0b5a6e3a71b59be1c7be006f0cfaaf508bde6ac07c811391: Status 404 returned error can't find the container with id d4b21179521e1a0c0b5a6e3a71b59be1c7be006f0cfaaf508bde6ac07c811391 Oct 11 04:04:49 crc kubenswrapper[4967]: E1011 04:04:49.844343 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a17fc270857869fd1efe5020b2a1cb8c2abbd838f08de88f3a6a59e8754ec351,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:true,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-baremetal-operator-agent:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_ANSIBLEEE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-ansibleee-runner:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_EVALUATOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-evaluator:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_AODH_NOTIFIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-aodh-notifier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_APACHE_IMAGE_URL_DEFAULT,Value:registry.redhat.io/ubi9/httpd-24:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_KEYSTONE_LISTENER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-keystone-listener:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_BARBICAN_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-barbican-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_IPMI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-ipmi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_MYSQLD_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/mysqld-exporter:v0.15.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_NOTIFICATION_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ceilometer-notification:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CEILOMETER_SGCORE_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/sg-core:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_BACKUP_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-backup:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_CINDER_VOLUME_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cinder-volume:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_BACKENDBIND9_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-backend-bind9:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_CENTRAL_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-central:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_MDNS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-mdns:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_PRODUCER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-producer:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_UNBOUND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-unbound:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_DESIGNATE_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-designate-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_FRR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-frr:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_ISCSID_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-iscsid:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_KEPLER_IMAGE_URL_DEFAULT,Value:quay.io/sustainable_computing_io/kepler:release-0.7.12,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_LOGROTATE_CROND_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-cron:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_MULTIPATHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-multipathd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_DHCP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-dhcp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_METADATA_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-metadata-agent-ovn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_OVN_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-ovn-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NEUTRON_SRIOV_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-sriov-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_NODE_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/prometheus/node-exporter:v1.5.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_OVN_BGP_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-bgp-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_EDPM_PODMAN_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/navidys/prometheus-podman-exporter:v1.10.1,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_GLANCE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-glance-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_CFNAPI_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-api-cfn:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HEAT_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-heat-engine:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_HORIZON_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-horizon:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_MEMCACHED_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-memcached:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_INFRA_REDIS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-redis:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_INSPECTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-inspector:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_NEUTRON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-neutron-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PXE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ironic-pxe:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_IRONIC_PYTHON_AGENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/ironic-python-agent:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KEYSTONE_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-keystone:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_KSM_IMAGE_URL_DEFAULT,Value:registry.k8s.io/kube-state-metrics/kube-state-metrics:v2.15.0,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MANILA_SHARE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-manila-share:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_MARIADB_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-mariadb:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NET_UTILS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-netutils:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NEUTRON_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_COMPUTE_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-compute:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_CONDUCTOR_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-conductor:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_NOVNC_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-novncproxy:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_NOVA_SCHEDULER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-nova-scheduler:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HEALTHMANAGER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-health-manager:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_HOUSEKEEPING_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-housekeeping:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_RSYSLOG_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rsyslog:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OCTAVIA_WORKER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-octavia-worker:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_CLIENT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-openstackclient:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_MUST_GATHER_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-must-gather:latest,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OPENSTACK_NETWORK_EXPORTER_IMAGE_URL_DEFAULT,Value:quay.io/openstack-k8s-operators/openstack-network-exporter:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OS_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/edpm-hardened-uefi:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-controller:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_CONTROLLER_OVS_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-base:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-nb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_NORTHD_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-northd:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_OVN_SB_DBCLUSTER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-ovn-sb-db-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_PLACEMENT_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-placement-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_RABBITMQ_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-rabbitmq:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_ACCOUNT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-account:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_CONTAINER_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-container:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_OBJECT_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-object:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_SWIFT_PROXY_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-swift-proxy-server:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_TEST_TEMPEST_IMAGE_URL_DEFAULT,Value:quay.io/podified-antelope-centos9/openstack-tempest-all:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_API_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-api:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_APPLIER_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-applier:current-podified,ValueFrom:nil,},EnvVar{Name:RELATED_IMAGE_WATCHER_DECISION_ENGINE_IMAGE_URL_DEFAULT,Value:quay.io/podified-master-centos9/openstack-watcher-decision-engine:current-podified,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:cert,ReadOnly:true,MountPath:/tmp/k8s-webhook-server/serving-certs,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-hql9t,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns_openstack-operators(0cc11444-d341-4821-a743-6b4af1466e4a): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 11 04:04:49 crc kubenswrapper[4967]: E1011 04:04:49.863425 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:operator,Image:quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2,Command:[/manager],Args:[],WorkingDir:,Ports:[]ContainerPort{ContainerPort{Name:metrics,HostPort:0,ContainerPort:9782,Protocol:TCP,HostIP:,},},Env:[]EnvVar{EnvVar{Name:OPERATOR_NAMESPACE,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:metadata.namespace,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{200 -3} {} 200m DecimalSI},memory: {{524288000 0} {} 500Mi BinarySI},},Requests:ResourceList{cpu: {{5 -3} {} 5m DecimalSI},memory: {{67108864 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-cpwn8,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000660000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod rabbitmq-cluster-operator-manager-5f97d8c699-cmzk2_openstack-operators(11e4f97c-6107-45fd-8eb1-e39ef7bed4ce): ErrImagePull: pull QPS exceeded" logger="UnhandledError" Oct 11 04:04:49 crc kubenswrapper[4967]: E1011 04:04:49.864735 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-cmzk2" podUID="11e4f97c-6107-45fd-8eb1-e39ef7bed4ce" Oct 11 04:04:50 crc kubenswrapper[4967]: E1011 04:04:50.153971 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/placement-operator-controller-manager-664664cb68-vng4c" podUID="26b9d5a0-dfbe-4e53-aa03-4fbb4682019b" Oct 11 04:04:50 crc kubenswrapper[4967]: E1011 04:04:50.167620 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/watcher-operator-controller-manager-646675d848-zdpbn" podUID="10a0ea36-9248-4086-83e4-de8578492421" Oct 11 04:04:50 crc kubenswrapper[4967]: E1011 04:04:50.167902 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns" podUID="0cc11444-d341-4821-a743-6b4af1466e4a" Oct 11 04:04:50 crc kubenswrapper[4967]: E1011 04:04:50.170015 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-4s7hv" podUID="5e24d70e-8882-45c0-9f6b-e443549ecabb" Oct 11 04:04:50 crc kubenswrapper[4967]: I1011 04:04:50.467109 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-66664b49c9-tk2lg" event={"ID":"620877eb-6f9c-4647-b545-1cca7a438690","Type":"ContainerStarted","Data":"9b30a12c127915cec43a5046c277114291467eb194afb2e31ba67ec2bf13e2a8"} Oct 11 04:04:50 crc kubenswrapper[4967]: I1011 04:04:50.468635 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-65knj" event={"ID":"1094a8be-16bb-4986-b4a4-465d41880800","Type":"ContainerStarted","Data":"182891a6f5158a0fbe8e9dad5d953309eae6da2178ea3d81d41e30d4d5ce6cf7"} Oct 11 04:04:50 crc kubenswrapper[4967]: I1011 04:04:50.473987 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-6zmv5" event={"ID":"538ebbcb-d51c-472c-a2b4-80236b0c8c08","Type":"ContainerStarted","Data":"f8a7df1ea72b2ece3832710ebf0ab1c02cf4f9b02eeb2948ac82b7c7fc711b30"} Oct 11 04:04:50 crc kubenswrapper[4967]: I1011 04:04:50.479309 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-685c45897-5gbwb" event={"ID":"6628a0ae-2a55-48b0-a5f1-0439c7d0c273","Type":"ContainerStarted","Data":"4f35e1e172151d7b8ab105b695066b3f743b0641ca31cf2332205442f72ce0e2"} Oct 11 04:04:50 crc kubenswrapper[4967]: I1011 04:04:50.508564 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns" event={"ID":"0cc11444-d341-4821-a743-6b4af1466e4a","Type":"ContainerStarted","Data":"232c9afbec16f225c34536edd8fbd2d85fa3b3def8b4e46bc2d87c72afa51c85"} Oct 11 04:04:50 crc kubenswrapper[4967]: I1011 04:04:50.508622 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns" event={"ID":"0cc11444-d341-4821-a743-6b4af1466e4a","Type":"ContainerStarted","Data":"9089b2f32b207344bccf3e939d8aafe58cf76491363672b10f2d39c7b4823e1f"} Oct 11 04:04:50 crc kubenswrapper[4967]: E1011 04:04:50.514241 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a17fc270857869fd1efe5020b2a1cb8c2abbd838f08de88f3a6a59e8754ec351\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns" podUID="0cc11444-d341-4821-a743-6b4af1466e4a" Oct 11 04:04:50 crc kubenswrapper[4967]: I1011 04:04:50.524831 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-fgz82" event={"ID":"e82102ad-391b-4b29-ae47-dae18dba7743","Type":"ContainerStarted","Data":"3a8ba1d19b7c124632fc6198595d99a0bd5446b98706e500023e92475d8730ec"} Oct 11 04:04:50 crc kubenswrapper[4967]: I1011 04:04:50.557415 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-646675d848-zdpbn" event={"ID":"10a0ea36-9248-4086-83e4-de8578492421","Type":"ContainerStarted","Data":"1ce95e67b7912e730841764415d34342907b26fb0cc26e0d7e9dccc73220bfff"} Oct 11 04:04:50 crc kubenswrapper[4967]: I1011 04:04:50.557468 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-646675d848-zdpbn" event={"ID":"10a0ea36-9248-4086-83e4-de8578492421","Type":"ContainerStarted","Data":"fccba244efb965e21ea982f9f9058efcbd07e08bb723fa4f4fb64dc750a1957a"} Oct 11 04:04:50 crc kubenswrapper[4967]: I1011 04:04:50.560188 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-cmzk2" event={"ID":"11e4f97c-6107-45fd-8eb1-e39ef7bed4ce","Type":"ContainerStarted","Data":"d4b21179521e1a0c0b5a6e3a71b59be1c7be006f0cfaaf508bde6ac07c811391"} Oct 11 04:04:50 crc kubenswrapper[4967]: E1011 04:04:50.560873 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:98a5233f0596591acdf2c6a5838b08be108787cdb6ad1995b2b7886bac0fe6ca\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-646675d848-zdpbn" podUID="10a0ea36-9248-4086-83e4-de8578492421" Oct 11 04:04:50 crc kubenswrapper[4967]: E1011 04:04:50.563275 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-cmzk2" podUID="11e4f97c-6107-45fd-8eb1-e39ef7bed4ce" Oct 11 04:04:50 crc kubenswrapper[4967]: I1011 04:04:50.565428 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-4s7hv" event={"ID":"5e24d70e-8882-45c0-9f6b-e443549ecabb","Type":"ContainerStarted","Data":"8c85a154bee5532b7dc1d11a8479576468f6e5817bcf0b5787e5f0cb1755fbec"} Oct 11 04:04:50 crc kubenswrapper[4967]: I1011 04:04:50.565465 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-4s7hv" event={"ID":"5e24d70e-8882-45c0-9f6b-e443549ecabb","Type":"ContainerStarted","Data":"f8d2485609cacc1daa4ba649a1d74ca6a1feae032a2b0006a3d1fa504400031e"} Oct 11 04:04:50 crc kubenswrapper[4967]: E1011 04:04:50.566836 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:7e584b1c430441c8b6591dadeff32e065de8a185ad37ef90d2e08d37e59aab4a\\\"\"" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-4s7hv" podUID="5e24d70e-8882-45c0-9f6b-e443549ecabb" Oct 11 04:04:50 crc kubenswrapper[4967]: I1011 04:04:50.570801 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-664664cb68-vng4c" event={"ID":"26b9d5a0-dfbe-4e53-aa03-4fbb4682019b","Type":"ContainerStarted","Data":"2224af30e4c2d9077471ca9b960a387c4085afd3b4c3d7670efb685bec078c66"} Oct 11 04:04:50 crc kubenswrapper[4967]: I1011 04:04:50.570828 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-664664cb68-vng4c" event={"ID":"26b9d5a0-dfbe-4e53-aa03-4fbb4682019b","Type":"ContainerStarted","Data":"ec812b0946ab5bab24e13f55a5fdd12b7d10f4c08c501325a64021f7d2ed70fc"} Oct 11 04:04:50 crc kubenswrapper[4967]: E1011 04:04:50.571441 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff\\\"\"" pod="openstack-operators/placement-operator-controller-manager-664664cb68-vng4c" podUID="26b9d5a0-dfbe-4e53-aa03-4fbb4682019b" Oct 11 04:04:51 crc kubenswrapper[4967]: I1011 04:04:51.306970 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8nwxv"] Oct 11 04:04:51 crc kubenswrapper[4967]: I1011 04:04:51.579336 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-8nwxv" podUID="8487ea91-dfea-423a-8c19-358a235b9eff" containerName="registry-server" containerID="cri-o://64e21d0eed5b8671f1f847782f3177afa710a544e25e1713597cfa9e2ce9d877" gracePeriod=2 Oct 11 04:04:51 crc kubenswrapper[4967]: E1011 04:04:51.582314 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/openstack-baremetal-operator@sha256:a17fc270857869fd1efe5020b2a1cb8c2abbd838f08de88f3a6a59e8754ec351\\\"\"" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns" podUID="0cc11444-d341-4821-a743-6b4af1466e4a" Oct 11 04:04:51 crc kubenswrapper[4967]: E1011 04:04:51.582928 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/test-operator@sha256:7e584b1c430441c8b6591dadeff32e065de8a185ad37ef90d2e08d37e59aab4a\\\"\"" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-4s7hv" podUID="5e24d70e-8882-45c0-9f6b-e443549ecabb" Oct 11 04:04:51 crc kubenswrapper[4967]: E1011 04:04:51.583034 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/watcher-operator@sha256:98a5233f0596591acdf2c6a5838b08be108787cdb6ad1995b2b7886bac0fe6ca\\\"\"" pod="openstack-operators/watcher-operator-controller-manager-646675d848-zdpbn" podUID="10a0ea36-9248-4086-83e4-de8578492421" Oct 11 04:04:51 crc kubenswrapper[4967]: E1011 04:04:51.583079 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"operator\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/rabbitmq-cluster-operator@sha256:893e66303c1b0bc1d00a299a3f0380bad55c8dc813c8a1c6a4aab379f5aa12a2\\\"\"" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-cmzk2" podUID="11e4f97c-6107-45fd-8eb1-e39ef7bed4ce" Oct 11 04:04:51 crc kubenswrapper[4967]: E1011 04:04:51.583133 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/placement-operator@sha256:d33c1f507e1f5b9a4bf226ad98917e92101ac66b36e19d35cbe04ae7014f6bff\\\"\"" pod="openstack-operators/placement-operator-controller-manager-664664cb68-vng4c" podUID="26b9d5a0-dfbe-4e53-aa03-4fbb4682019b" Oct 11 04:04:52 crc kubenswrapper[4967]: I1011 04:04:52.588315 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z9scd" event={"ID":"7de04fa6-4dd6-4254-bde0-b83b920be0fa","Type":"ContainerStarted","Data":"d02ce1fa87dc4a67713923b2f5f5d220839cbf12d7ef7eb22c9bd771fb3c8920"} Oct 11 04:04:52 crc kubenswrapper[4967]: I1011 04:04:52.589305 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-66664b49c9-tk2lg" event={"ID":"620877eb-6f9c-4647-b545-1cca7a438690","Type":"ContainerStarted","Data":"cc3e82af623dc9a41d8a687d2e6b108d81ce0a340ad82ae95ada3f2750900eba"} Oct 11 04:04:52 crc kubenswrapper[4967]: I1011 04:04:52.607739 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-z9scd" podStartSLOduration=4.110501882 podStartE2EDuration="7.607720429s" podCreationTimestamp="2025-10-11 04:04:45 +0000 UTC" firstStartedPulling="2025-10-11 04:04:46.390051897 +0000 UTC m=+814.353260830" lastFinishedPulling="2025-10-11 04:04:49.887270444 +0000 UTC m=+817.850479377" observedRunningTime="2025-10-11 04:04:52.60227479 +0000 UTC m=+820.565483743" watchObservedRunningTime="2025-10-11 04:04:52.607720429 +0000 UTC m=+820.570929372" Oct 11 04:04:53 crc kubenswrapper[4967]: E1011 04:04:53.346695 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-fgz82" podUID="e82102ad-391b-4b29-ae47-dae18dba7743" Oct 11 04:04:53 crc kubenswrapper[4967]: I1011 04:04:53.598319 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-fgz82" event={"ID":"e82102ad-391b-4b29-ae47-dae18dba7743","Type":"ContainerStarted","Data":"75d976610fcd7a0ff9ed50b5b9460403daccf65bcca5f1c496f573d43804bd53"} Oct 11 04:04:53 crc kubenswrapper[4967]: E1011 04:04:53.599718 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:abe978f8da75223de5043cca50278ad4e28c8dd309883f502fe1e7a9998733b0\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-fgz82" podUID="e82102ad-391b-4b29-ae47-dae18dba7743" Oct 11 04:04:53 crc kubenswrapper[4967]: I1011 04:04:53.601811 4967 generic.go:334] "Generic (PLEG): container finished" podID="8487ea91-dfea-423a-8c19-358a235b9eff" containerID="64e21d0eed5b8671f1f847782f3177afa710a544e25e1713597cfa9e2ce9d877" exitCode=0 Oct 11 04:04:53 crc kubenswrapper[4967]: I1011 04:04:53.601897 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8nwxv" event={"ID":"8487ea91-dfea-423a-8c19-358a235b9eff","Type":"ContainerDied","Data":"64e21d0eed5b8671f1f847782f3177afa710a544e25e1713597cfa9e2ce9d877"} Oct 11 04:04:54 crc kubenswrapper[4967]: E1011 04:04:54.609914 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/telemetry-operator@sha256:abe978f8da75223de5043cca50278ad4e28c8dd309883f502fe1e7a9998733b0\\\"\"" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-fgz82" podUID="e82102ad-391b-4b29-ae47-dae18dba7743" Oct 11 04:04:55 crc kubenswrapper[4967]: I1011 04:04:55.627846 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-z9scd" Oct 11 04:04:55 crc kubenswrapper[4967]: I1011 04:04:55.628238 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-z9scd" Oct 11 04:04:55 crc kubenswrapper[4967]: I1011 04:04:55.630281 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-8nwxv" event={"ID":"8487ea91-dfea-423a-8c19-358a235b9eff","Type":"ContainerDied","Data":"c067624ff054a45efe5dcf20c045d1ed3750a1a91365a4a142a8fc8994993538"} Oct 11 04:04:55 crc kubenswrapper[4967]: I1011 04:04:55.630329 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c067624ff054a45efe5dcf20c045d1ed3750a1a91365a4a142a8fc8994993538" Oct 11 04:04:55 crc kubenswrapper[4967]: I1011 04:04:55.693641 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-z9scd" Oct 11 04:04:55 crc kubenswrapper[4967]: I1011 04:04:55.747330 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8nwxv" Oct 11 04:04:55 crc kubenswrapper[4967]: E1011 04:04:55.816499 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"pull QPS exceeded\"" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-65knj" podUID="1094a8be-16bb-4986-b4a4-465d41880800" Oct 11 04:04:55 crc kubenswrapper[4967]: I1011 04:04:55.851489 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xpvw8\" (UniqueName: \"kubernetes.io/projected/8487ea91-dfea-423a-8c19-358a235b9eff-kube-api-access-xpvw8\") pod \"8487ea91-dfea-423a-8c19-358a235b9eff\" (UID: \"8487ea91-dfea-423a-8c19-358a235b9eff\") " Oct 11 04:04:55 crc kubenswrapper[4967]: I1011 04:04:55.851629 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8487ea91-dfea-423a-8c19-358a235b9eff-catalog-content\") pod \"8487ea91-dfea-423a-8c19-358a235b9eff\" (UID: \"8487ea91-dfea-423a-8c19-358a235b9eff\") " Oct 11 04:04:55 crc kubenswrapper[4967]: I1011 04:04:55.851742 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8487ea91-dfea-423a-8c19-358a235b9eff-utilities\") pod \"8487ea91-dfea-423a-8c19-358a235b9eff\" (UID: \"8487ea91-dfea-423a-8c19-358a235b9eff\") " Oct 11 04:04:55 crc kubenswrapper[4967]: I1011 04:04:55.853165 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8487ea91-dfea-423a-8c19-358a235b9eff-utilities" (OuterVolumeSpecName: "utilities") pod "8487ea91-dfea-423a-8c19-358a235b9eff" (UID: "8487ea91-dfea-423a-8c19-358a235b9eff"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:04:55 crc kubenswrapper[4967]: I1011 04:04:55.859338 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8487ea91-dfea-423a-8c19-358a235b9eff-kube-api-access-xpvw8" (OuterVolumeSpecName: "kube-api-access-xpvw8") pod "8487ea91-dfea-423a-8c19-358a235b9eff" (UID: "8487ea91-dfea-423a-8c19-358a235b9eff"). InnerVolumeSpecName "kube-api-access-xpvw8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:04:55 crc kubenswrapper[4967]: I1011 04:04:55.930403 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8487ea91-dfea-423a-8c19-358a235b9eff-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8487ea91-dfea-423a-8c19-358a235b9eff" (UID: "8487ea91-dfea-423a-8c19-358a235b9eff"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:04:55 crc kubenswrapper[4967]: I1011 04:04:55.953523 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8487ea91-dfea-423a-8c19-358a235b9eff-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 04:04:55 crc kubenswrapper[4967]: I1011 04:04:55.953564 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8487ea91-dfea-423a-8c19-358a235b9eff-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 04:04:55 crc kubenswrapper[4967]: I1011 04:04:55.953578 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xpvw8\" (UniqueName: \"kubernetes.io/projected/8487ea91-dfea-423a-8c19-358a235b9eff-kube-api-access-xpvw8\") on node \"crc\" DevicePath \"\"" Oct 11 04:04:56 crc kubenswrapper[4967]: I1011 04:04:56.638842 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-8nwxv" Oct 11 04:04:56 crc kubenswrapper[4967]: I1011 04:04:56.638837 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-65knj" event={"ID":"1094a8be-16bb-4986-b4a4-465d41880800","Type":"ContainerStarted","Data":"5043cc0536dbf2eb13e9bfedf8bf1834858ef9cc3847c4ba806327b897f54f07"} Oct 11 04:04:56 crc kubenswrapper[4967]: E1011 04:04:56.642214 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e\\\"\"" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-65knj" podUID="1094a8be-16bb-4986-b4a4-465d41880800" Oct 11 04:04:56 crc kubenswrapper[4967]: I1011 04:04:56.724220 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-z9scd" Oct 11 04:04:56 crc kubenswrapper[4967]: I1011 04:04:56.725493 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-8nwxv"] Oct 11 04:04:56 crc kubenswrapper[4967]: I1011 04:04:56.731612 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-8nwxv"] Oct 11 04:04:56 crc kubenswrapper[4967]: I1011 04:04:56.826927 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8487ea91-dfea-423a-8c19-358a235b9eff" path="/var/lib/kubelet/pods/8487ea91-dfea-423a-8c19-358a235b9eff/volumes" Oct 11 04:04:57 crc kubenswrapper[4967]: E1011 04:04:57.649847 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/swift-operator@sha256:4b4a17fe08ce00e375afaaec6a28835f5c1784f03d11c4558376ac04130f3a9e\\\"\"" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-65knj" podUID="1094a8be-16bb-4986-b4a4-465d41880800" Oct 11 04:04:58 crc kubenswrapper[4967]: I1011 04:04:58.931196 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z9scd"] Oct 11 04:04:58 crc kubenswrapper[4967]: I1011 04:04:58.931487 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-z9scd" podUID="7de04fa6-4dd6-4254-bde0-b83b920be0fa" containerName="registry-server" containerID="cri-o://d02ce1fa87dc4a67713923b2f5f5d220839cbf12d7ef7eb22c9bd771fb3c8920" gracePeriod=2 Oct 11 04:05:00 crc kubenswrapper[4967]: I1011 04:05:00.677964 4967 generic.go:334] "Generic (PLEG): container finished" podID="7de04fa6-4dd6-4254-bde0-b83b920be0fa" containerID="d02ce1fa87dc4a67713923b2f5f5d220839cbf12d7ef7eb22c9bd771fb3c8920" exitCode=0 Oct 11 04:05:00 crc kubenswrapper[4967]: I1011 04:05:00.678025 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z9scd" event={"ID":"7de04fa6-4dd6-4254-bde0-b83b920be0fa","Type":"ContainerDied","Data":"d02ce1fa87dc4a67713923b2f5f5d220839cbf12d7ef7eb22c9bd771fb3c8920"} Oct 11 04:05:01 crc kubenswrapper[4967]: E1011 04:05:01.661178 4967 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/barbican-operator@sha256:783f711b4cb179819cfcb81167c3591c70671440f4551bbe48b7a8730567f577" Oct 11 04:05:01 crc kubenswrapper[4967]: E1011 04:05:01.661400 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/barbican-operator@sha256:783f711b4cb179819cfcb81167c3591c70671440f4551bbe48b7a8730567f577,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-9kxq9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod barbican-operator-controller-manager-64f84fcdbb-xz92m_openstack-operators(6f3fbb7d-1207-45a2-8213-c7c71bd703f1): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 11 04:05:02 crc kubenswrapper[4967]: E1011 04:05:02.785803 4967 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/mariadb-operator@sha256:47278ed28e02df00892f941763aa0d69547327318e8a983e07f4577acd288167" Oct 11 04:05:02 crc kubenswrapper[4967]: E1011 04:05:02.786050 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/mariadb-operator@sha256:47278ed28e02df00892f941763aa0d69547327318e8a983e07f4577acd288167,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-tgnl4,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod mariadb-operator-controller-manager-5777b4f897-fjr8k_openstack-operators(ef8ae8f4-ad95-4f1f-bc8c-170fe5058dc9): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 11 04:05:03 crc kubenswrapper[4967]: E1011 04:05:03.385142 4967 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/heat-operator@sha256:ec11cb8711bd1af22db3c84aa854349ee46191add3db45aecfabb1d8410c04d0" Oct 11 04:05:03 crc kubenswrapper[4967]: E1011 04:05:03.385320 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/heat-operator@sha256:ec11cb8711bd1af22db3c84aa854349ee46191add3db45aecfabb1d8410c04d0,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-bcq5z,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod heat-operator-controller-manager-6d9967f8dd-ww7nt_openstack-operators(ecd905f3-adb2-47ae-a86d-6fc0b4907a76): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 11 04:05:04 crc kubenswrapper[4967]: E1011 04:05:04.748040 4967 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/nova-operator@sha256:b2e9acf568a48c28cf2aed6012e432eeeb7d5f0eb11878fc91b62bc34cba10cd" Oct 11 04:05:04 crc kubenswrapper[4967]: E1011 04:05:04.748277 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/nova-operator@sha256:b2e9acf568a48c28cf2aed6012e432eeeb7d5f0eb11878fc91b62bc34cba10cd,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-ldjzg,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod nova-operator-controller-manager-57bb74c7bf-59px8_openstack-operators(7140d0e1-9937-491b-aa8a-16db830aff11): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 11 04:05:05 crc kubenswrapper[4967]: E1011 04:05:05.264612 4967 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/openstack-k8s-operators/keystone-operator@sha256:79b43a69884631c635d2164b95a2d4ec68f5cb33f96da14764f1c710880f3997" Oct 11 04:05:05 crc kubenswrapper[4967]: E1011 04:05:05.264823 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:manager,Image:quay.io/openstack-k8s-operators/keystone-operator@sha256:79b43a69884631c635d2164b95a2d4ec68f5cb33f96da14764f1c710880f3997,Command:[/manager],Args:[--health-probe-bind-address=:8081 --metrics-bind-address=127.0.0.1:8080 --leader-elect],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:LEASE_DURATION,Value:30,ValueFrom:nil,},EnvVar{Name:RENEW_DEADLINE,Value:20,ValueFrom:nil,},EnvVar{Name:RETRY_PERIOD,Value:5,ValueFrom:nil,},EnvVar{Name:ENABLE_WEBHOOKS,Value:false,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{cpu: {{500 -3} {} 500m DecimalSI},memory: {{536870912 0} {} BinarySI},},Requests:ResourceList{cpu: {{10 -3} {} 10m DecimalSI},memory: {{268435456 0} {} BinarySI},},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-dbxkn,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/healthz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:15,TimeoutSeconds:1,PeriodSeconds:20,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},ReadinessProbe:&Probe{ProbeHandler:ProbeHandler{Exec:nil,HTTPGet:&HTTPGetAction{Path:/readyz,Port:{0 8081 },Host:,Scheme:HTTP,HTTPHeaders:[]HTTPHeader{},},TCPSocket:nil,GRPC:nil,},InitialDelaySeconds:5,TimeoutSeconds:1,PeriodSeconds:10,SuccessThreshold:1,FailureThreshold:3,TerminationGracePeriodSeconds:nil,},Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[MKNOD],},Privileged:nil,SELinuxOptions:nil,RunAsUser:nil,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod keystone-operator-controller-manager-ddb98f99b-brbrs_openstack-operators(950d3635-f160-45a1-9a4c-749720e662dc): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 11 04:05:05 crc kubenswrapper[4967]: E1011 04:05:05.629156 4967 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d02ce1fa87dc4a67713923b2f5f5d220839cbf12d7ef7eb22c9bd771fb3c8920 is running failed: container process not found" containerID="d02ce1fa87dc4a67713923b2f5f5d220839cbf12d7ef7eb22c9bd771fb3c8920" cmd=["grpc_health_probe","-addr=:50051"] Oct 11 04:05:05 crc kubenswrapper[4967]: E1011 04:05:05.629551 4967 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d02ce1fa87dc4a67713923b2f5f5d220839cbf12d7ef7eb22c9bd771fb3c8920 is running failed: container process not found" containerID="d02ce1fa87dc4a67713923b2f5f5d220839cbf12d7ef7eb22c9bd771fb3c8920" cmd=["grpc_health_probe","-addr=:50051"] Oct 11 04:05:05 crc kubenswrapper[4967]: E1011 04:05:05.629970 4967 log.go:32] "ExecSync cmd from runtime service failed" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d02ce1fa87dc4a67713923b2f5f5d220839cbf12d7ef7eb22c9bd771fb3c8920 is running failed: container process not found" containerID="d02ce1fa87dc4a67713923b2f5f5d220839cbf12d7ef7eb22c9bd771fb3c8920" cmd=["grpc_health_probe","-addr=:50051"] Oct 11 04:05:05 crc kubenswrapper[4967]: E1011 04:05:05.630093 4967 prober.go:104] "Probe errored" err="rpc error: code = NotFound desc = container is not created or running: checking if PID of d02ce1fa87dc4a67713923b2f5f5d220839cbf12d7ef7eb22c9bd771fb3c8920 is running failed: container process not found" probeType="Readiness" pod="openshift-marketplace/redhat-marketplace-z9scd" podUID="7de04fa6-4dd6-4254-bde0-b83b920be0fa" containerName="registry-server" Oct 11 04:05:05 crc kubenswrapper[4967]: I1011 04:05:05.717415 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-z9scd" event={"ID":"7de04fa6-4dd6-4254-bde0-b83b920be0fa","Type":"ContainerDied","Data":"148bab79652295b728efd895d90bbd1c9011a5108116cdd9cf94d7f264961b3a"} Oct 11 04:05:05 crc kubenswrapper[4967]: I1011 04:05:05.717457 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="148bab79652295b728efd895d90bbd1c9011a5108116cdd9cf94d7f264961b3a" Oct 11 04:05:05 crc kubenswrapper[4967]: I1011 04:05:05.747216 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z9scd" Oct 11 04:05:05 crc kubenswrapper[4967]: I1011 04:05:05.820347 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7de04fa6-4dd6-4254-bde0-b83b920be0fa-catalog-content\") pod \"7de04fa6-4dd6-4254-bde0-b83b920be0fa\" (UID: \"7de04fa6-4dd6-4254-bde0-b83b920be0fa\") " Oct 11 04:05:05 crc kubenswrapper[4967]: I1011 04:05:05.820720 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7de04fa6-4dd6-4254-bde0-b83b920be0fa-utilities\") pod \"7de04fa6-4dd6-4254-bde0-b83b920be0fa\" (UID: \"7de04fa6-4dd6-4254-bde0-b83b920be0fa\") " Oct 11 04:05:05 crc kubenswrapper[4967]: I1011 04:05:05.820745 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-96b89\" (UniqueName: \"kubernetes.io/projected/7de04fa6-4dd6-4254-bde0-b83b920be0fa-kube-api-access-96b89\") pod \"7de04fa6-4dd6-4254-bde0-b83b920be0fa\" (UID: \"7de04fa6-4dd6-4254-bde0-b83b920be0fa\") " Oct 11 04:05:05 crc kubenswrapper[4967]: I1011 04:05:05.825665 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7de04fa6-4dd6-4254-bde0-b83b920be0fa-utilities" (OuterVolumeSpecName: "utilities") pod "7de04fa6-4dd6-4254-bde0-b83b920be0fa" (UID: "7de04fa6-4dd6-4254-bde0-b83b920be0fa"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:05:05 crc kubenswrapper[4967]: I1011 04:05:05.828159 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7de04fa6-4dd6-4254-bde0-b83b920be0fa-kube-api-access-96b89" (OuterVolumeSpecName: "kube-api-access-96b89") pod "7de04fa6-4dd6-4254-bde0-b83b920be0fa" (UID: "7de04fa6-4dd6-4254-bde0-b83b920be0fa"). InnerVolumeSpecName "kube-api-access-96b89". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:05:05 crc kubenswrapper[4967]: I1011 04:05:05.833941 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/7de04fa6-4dd6-4254-bde0-b83b920be0fa-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "7de04fa6-4dd6-4254-bde0-b83b920be0fa" (UID: "7de04fa6-4dd6-4254-bde0-b83b920be0fa"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:05:05 crc kubenswrapper[4967]: I1011 04:05:05.922348 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/7de04fa6-4dd6-4254-bde0-b83b920be0fa-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 04:05:05 crc kubenswrapper[4967]: I1011 04:05:05.922400 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/7de04fa6-4dd6-4254-bde0-b83b920be0fa-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 04:05:05 crc kubenswrapper[4967]: I1011 04:05:05.922413 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-96b89\" (UniqueName: \"kubernetes.io/projected/7de04fa6-4dd6-4254-bde0-b83b920be0fa-kube-api-access-96b89\") on node \"crc\" DevicePath \"\"" Oct 11 04:05:06 crc kubenswrapper[4967]: I1011 04:05:06.722596 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-z9scd" Oct 11 04:05:06 crc kubenswrapper[4967]: I1011 04:05:06.778172 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-z9scd"] Oct 11 04:05:06 crc kubenswrapper[4967]: I1011 04:05:06.784598 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-z9scd"] Oct 11 04:05:06 crc kubenswrapper[4967]: I1011 04:05:06.827324 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7de04fa6-4dd6-4254-bde0-b83b920be0fa" path="/var/lib/kubelet/pods/7de04fa6-4dd6-4254-bde0-b83b920be0fa/volumes" Oct 11 04:05:06 crc kubenswrapper[4967]: E1011 04:05:06.846185 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-fjr8k" podUID="ef8ae8f4-ad95-4f1f-bc8c-170fe5058dc9" Oct 11 04:05:07 crc kubenswrapper[4967]: I1011 04:05:07.731338 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-operator-controller-manager-66664b49c9-tk2lg" event={"ID":"620877eb-6f9c-4647-b545-1cca7a438690","Type":"ContainerStarted","Data":"408a5e2fa410eb8a7d7ac6c20cee058ae3742ed9a6494ae9bdd926689a057252"} Oct 11 04:05:07 crc kubenswrapper[4967]: I1011 04:05:07.732492 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-operator-controller-manager-66664b49c9-tk2lg" Oct 11 04:05:07 crc kubenswrapper[4967]: I1011 04:05:07.734710 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-fjr8k" event={"ID":"ef8ae8f4-ad95-4f1f-bc8c-170fe5058dc9","Type":"ContainerStarted","Data":"2204aac27bb55d6e650da6f739809255e00defa88557922fa86380f77c5f6cdd"} Oct 11 04:05:07 crc kubenswrapper[4967]: I1011 04:05:07.737638 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-operator-controller-manager-66664b49c9-tk2lg" Oct 11 04:05:07 crc kubenswrapper[4967]: E1011 04:05:07.741971 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:47278ed28e02df00892f941763aa0d69547327318e8a983e07f4577acd288167\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-fjr8k" podUID="ef8ae8f4-ad95-4f1f-bc8c-170fe5058dc9" Oct 11 04:05:07 crc kubenswrapper[4967]: I1011 04:05:07.765358 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-operator-controller-manager-66664b49c9-tk2lg" podStartSLOduration=19.765341861 podStartE2EDuration="19.765341861s" podCreationTimestamp="2025-10-11 04:04:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:05:07.759146804 +0000 UTC m=+835.722355757" watchObservedRunningTime="2025-10-11 04:05:07.765341861 +0000 UTC m=+835.728550794" Oct 11 04:05:07 crc kubenswrapper[4967]: E1011 04:05:07.875404 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-xz92m" podUID="6f3fbb7d-1207-45a2-8213-c7c71bd703f1" Oct 11 04:05:07 crc kubenswrapper[4967]: E1011 04:05:07.876710 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-ww7nt" podUID="ecd905f3-adb2-47ae-a86d-6fc0b4907a76" Oct 11 04:05:08 crc kubenswrapper[4967]: E1011 04:05:08.185635 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-brbrs" podUID="950d3635-f160-45a1-9a4c-749720e662dc" Oct 11 04:05:08 crc kubenswrapper[4967]: E1011 04:05:08.197570 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-59px8" podUID="7140d0e1-9937-491b-aa8a-16db830aff11" Oct 11 04:05:08 crc kubenswrapper[4967]: I1011 04:05:08.806304 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-kkrkc" event={"ID":"c9e29e65-cdd1-408d-bed0-bc283147b283","Type":"ContainerStarted","Data":"94cc756b5215ef478d2ab57d53225a5f91e41e4a2bc73c6c937665fe00b22d96"} Oct 11 04:05:08 crc kubenswrapper[4967]: I1011 04:05:08.827875 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-9gn9l" event={"ID":"2a2805ea-7d92-42e4-89cb-9b2672f72fc6","Type":"ContainerStarted","Data":"b8af77a2d5431b64f4f91c17053eacae1c6cd25a721ca1e1b49908bfb7db0275"} Oct 11 04:05:08 crc kubenswrapper[4967]: I1011 04:05:08.847285 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-8ccrl" event={"ID":"e9feb54d-dbfa-40ec-ac45-535cca431024","Type":"ContainerStarted","Data":"c05b972f67d8064fdc3077a25545d9c11b2480860aa9d723e6d3d9b5ca7a410b"} Oct 11 04:05:08 crc kubenswrapper[4967]: I1011 04:05:08.865568 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-5glgp" event={"ID":"91f45294-2d2f-4b40-a1b4-a841c7748cc8","Type":"ContainerStarted","Data":"38de02e37a7767efda1bbc3207f909dbff2bcd32b0504923c88d37f090e1a734"} Oct 11 04:05:08 crc kubenswrapper[4967]: I1011 04:05:08.891114 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-59px8" event={"ID":"7140d0e1-9937-491b-aa8a-16db830aff11","Type":"ContainerStarted","Data":"1ade0d781fcef275642dbec53f67342d14d027f34be7cf828d82bcd21425ac7e"} Oct 11 04:05:08 crc kubenswrapper[4967]: E1011 04:05:08.901306 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:b2e9acf568a48c28cf2aed6012e432eeeb7d5f0eb11878fc91b62bc34cba10cd\\\"\"" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-59px8" podUID="7140d0e1-9937-491b-aa8a-16db830aff11" Oct 11 04:05:08 crc kubenswrapper[4967]: I1011 04:05:08.924735 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns" event={"ID":"0cc11444-d341-4821-a743-6b4af1466e4a","Type":"ContainerStarted","Data":"ff0df0b2cc962b4b44c2ffcb79943fdd166acad18e7b133cd267e5833275bf4c"} Oct 11 04:05:08 crc kubenswrapper[4967]: I1011 04:05:08.925462 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns" Oct 11 04:05:08 crc kubenswrapper[4967]: I1011 04:05:08.963044 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67494cb947-sklz6" event={"ID":"674852f1-9729-4a6e-8c90-31741fa69b49","Type":"ContainerStarted","Data":"87d2658a9524263cbf1c17d049aecb22c20508807ec8923a30f9b508da46426a"} Oct 11 04:05:08 crc kubenswrapper[4967]: I1011 04:05:08.987515 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/watcher-operator-controller-manager-646675d848-zdpbn" event={"ID":"10a0ea36-9248-4086-83e4-de8578492421","Type":"ContainerStarted","Data":"32f881badaefea4f114ddfdb9ae13bcbfb40cd78a97f73d0b984b7970b094e16"} Oct 11 04:05:08 crc kubenswrapper[4967]: I1011 04:05:08.988150 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/watcher-operator-controller-manager-646675d848-zdpbn" Oct 11 04:05:08 crc kubenswrapper[4967]: I1011 04:05:08.989682 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-brbrs" event={"ID":"950d3635-f160-45a1-9a4c-749720e662dc","Type":"ContainerStarted","Data":"4a9f78f2172770feba7cc3681b57e67769bcd542229553226b452b236b57f37a"} Oct 11 04:05:08 crc kubenswrapper[4967]: E1011 04:05:08.990692 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:79b43a69884631c635d2164b95a2d4ec68f5cb33f96da14764f1c710880f3997\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-brbrs" podUID="950d3635-f160-45a1-9a4c-749720e662dc" Oct 11 04:05:09 crc kubenswrapper[4967]: I1011 04:05:09.001469 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-7clg2" event={"ID":"7e917a6a-9720-4427-b079-33d112a981a3","Type":"ContainerStarted","Data":"ee5b9259df818cfc2e45fff3ef53618e5425355dfe78dce3b29da30ce2c6f0c0"} Oct 11 04:05:09 crc kubenswrapper[4967]: I1011 04:05:09.001511 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-7clg2" event={"ID":"7e917a6a-9720-4427-b079-33d112a981a3","Type":"ContainerStarted","Data":"76592640daed3645255d29aa67debd0d869db9072b7c9dee3d9e8762c019f2be"} Oct 11 04:05:09 crc kubenswrapper[4967]: I1011 04:05:09.002115 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-7clg2" Oct 11 04:05:09 crc kubenswrapper[4967]: I1011 04:05:09.019270 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-ww7nt" event={"ID":"ecd905f3-adb2-47ae-a86d-6fc0b4907a76","Type":"ContainerStarted","Data":"b3dd8d0f21442b38e49582e5a84e3aacd41ad8817cd7734529b4b1d90764b2d0"} Oct 11 04:05:09 crc kubenswrapper[4967]: E1011 04:05:09.023413 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:ec11cb8711bd1af22db3c84aa854349ee46191add3db45aecfabb1d8410c04d0\\\"\"" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-ww7nt" podUID="ecd905f3-adb2-47ae-a86d-6fc0b4907a76" Oct 11 04:05:09 crc kubenswrapper[4967]: I1011 04:05:09.046671 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-59578bc799-ls9k6" event={"ID":"20584c10-d8a1-41ca-a0a2-9465fa5190f2","Type":"ContainerStarted","Data":"5d942dd5f210a717e3d622098df2a81388e0ec274f90241ff61fc65cc7528e08"} Oct 11 04:05:09 crc kubenswrapper[4967]: I1011 04:05:09.046911 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/manila-operator-controller-manager-59578bc799-ls9k6" Oct 11 04:05:09 crc kubenswrapper[4967]: I1011 04:05:09.072946 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-6zmv5" event={"ID":"538ebbcb-d51c-472c-a2b4-80236b0c8c08","Type":"ContainerStarted","Data":"050471b5511325138842c8efbf780c84eac36b959abb41f71fe1b17f85b0cf1b"} Oct 11 04:05:09 crc kubenswrapper[4967]: I1011 04:05:09.091250 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-685c45897-5gbwb" event={"ID":"6628a0ae-2a55-48b0-a5f1-0439c7d0c273","Type":"ContainerStarted","Data":"c32e5c72f828e7703770d95c7e4e40ec3cde366d190362777ac91f08a06bf0be"} Oct 11 04:05:09 crc kubenswrapper[4967]: I1011 04:05:09.097315 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns" podStartSLOduration=3.971592635 podStartE2EDuration="22.09729194s" podCreationTimestamp="2025-10-11 04:04:47 +0000 UTC" firstStartedPulling="2025-10-11 04:04:49.843987661 +0000 UTC m=+817.807196594" lastFinishedPulling="2025-10-11 04:05:07.969686966 +0000 UTC m=+835.932895899" observedRunningTime="2025-10-11 04:05:09.091528444 +0000 UTC m=+837.054737377" watchObservedRunningTime="2025-10-11 04:05:09.09729194 +0000 UTC m=+837.060500883" Oct 11 04:05:09 crc kubenswrapper[4967]: I1011 04:05:09.101525 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-4s7hv" event={"ID":"5e24d70e-8882-45c0-9f6b-e443549ecabb","Type":"ContainerStarted","Data":"7c85f578bab862464792ef2d5dd8172ad5d7ef5753baad09329f145b81510064"} Oct 11 04:05:09 crc kubenswrapper[4967]: I1011 04:05:09.102219 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-4s7hv" Oct 11 04:05:09 crc kubenswrapper[4967]: I1011 04:05:09.129041 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-xz92m" event={"ID":"6f3fbb7d-1207-45a2-8213-c7c71bd703f1","Type":"ContainerStarted","Data":"04041c42d1fe7bcbd64dc15e579e9bc0c7ff3bd69e407320ebd0f73ade5d68c0"} Oct 11 04:05:09 crc kubenswrapper[4967]: E1011 04:05:09.130358 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/barbican-operator@sha256:783f711b4cb179819cfcb81167c3591c70671440f4551bbe48b7a8730567f577\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-xz92m" podUID="6f3fbb7d-1207-45a2-8213-c7c71bd703f1" Oct 11 04:05:09 crc kubenswrapper[4967]: I1011 04:05:09.134527 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-cmzk2" event={"ID":"11e4f97c-6107-45fd-8eb1-e39ef7bed4ce","Type":"ContainerStarted","Data":"a29c1339c1bd507130a7485b40adde820faba46bd243a0e6de5b3aecc035f804"} Oct 11 04:05:09 crc kubenswrapper[4967]: I1011 04:05:09.148872 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-zt44b" event={"ID":"44eb39fa-5b20-4c02-b039-f23d7c5220d2","Type":"ContainerStarted","Data":"9a0006b73591d91a0f853198d2fe0147c5e40400e8bbe7d85a58e7ce89b7afe6"} Oct 11 04:05:09 crc kubenswrapper[4967]: E1011 04:05:09.152845 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/mariadb-operator@sha256:47278ed28e02df00892f941763aa0d69547327318e8a983e07f4577acd288167\\\"\"" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-fjr8k" podUID="ef8ae8f4-ad95-4f1f-bc8c-170fe5058dc9" Oct 11 04:05:09 crc kubenswrapper[4967]: I1011 04:05:09.231168 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/watcher-operator-controller-manager-646675d848-zdpbn" podStartSLOduration=3.210747475 podStartE2EDuration="21.231147702s" podCreationTimestamp="2025-10-11 04:04:48 +0000 UTC" firstStartedPulling="2025-10-11 04:04:49.812779345 +0000 UTC m=+817.775988278" lastFinishedPulling="2025-10-11 04:05:07.833179572 +0000 UTC m=+835.796388505" observedRunningTime="2025-10-11 04:05:09.162262565 +0000 UTC m=+837.125471498" watchObservedRunningTime="2025-10-11 04:05:09.231147702 +0000 UTC m=+837.194356625" Oct 11 04:05:09 crc kubenswrapper[4967]: I1011 04:05:09.328949 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/manila-operator-controller-manager-59578bc799-ls9k6" podStartSLOduration=6.797734758 podStartE2EDuration="22.328934621s" podCreationTimestamp="2025-10-11 04:04:47 +0000 UTC" firstStartedPulling="2025-10-11 04:04:49.210734099 +0000 UTC m=+817.173943032" lastFinishedPulling="2025-10-11 04:05:04.741933932 +0000 UTC m=+832.705142895" observedRunningTime="2025-10-11 04:05:09.322753784 +0000 UTC m=+837.285962717" watchObservedRunningTime="2025-10-11 04:05:09.328934621 +0000 UTC m=+837.292143554" Oct 11 04:05:09 crc kubenswrapper[4967]: I1011 04:05:09.362372 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-7clg2" podStartSLOduration=6.042972416 podStartE2EDuration="22.362355689s" podCreationTimestamp="2025-10-11 04:04:47 +0000 UTC" firstStartedPulling="2025-10-11 04:04:49.36110634 +0000 UTC m=+817.324315273" lastFinishedPulling="2025-10-11 04:05:05.680489573 +0000 UTC m=+833.643698546" observedRunningTime="2025-10-11 04:05:09.354362981 +0000 UTC m=+837.317571914" watchObservedRunningTime="2025-10-11 04:05:09.362355689 +0000 UTC m=+837.325564622" Oct 11 04:05:09 crc kubenswrapper[4967]: I1011 04:05:09.422436 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/rabbitmq-cluster-operator-manager-5f97d8c699-cmzk2" podStartSLOduration=3.50915569 podStartE2EDuration="21.422421478s" podCreationTimestamp="2025-10-11 04:04:48 +0000 UTC" firstStartedPulling="2025-10-11 04:04:49.863284047 +0000 UTC m=+817.826492980" lastFinishedPulling="2025-10-11 04:05:07.776549835 +0000 UTC m=+835.739758768" observedRunningTime="2025-10-11 04:05:09.419339115 +0000 UTC m=+837.382548058" watchObservedRunningTime="2025-10-11 04:05:09.422421478 +0000 UTC m=+837.385630411" Oct 11 04:05:10 crc kubenswrapper[4967]: I1011 04:05:10.158761 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/infra-operator-controller-manager-685c45897-5gbwb" event={"ID":"6628a0ae-2a55-48b0-a5f1-0439c7d0c273","Type":"ContainerStarted","Data":"cae7c766453e25b595883f1a6fcc97d45ddbc6825b8cfd036a018a1e94bf3d8d"} Oct 11 04:05:10 crc kubenswrapper[4967]: I1011 04:05:10.159057 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/infra-operator-controller-manager-685c45897-5gbwb" Oct 11 04:05:10 crc kubenswrapper[4967]: I1011 04:05:10.161830 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-zt44b" event={"ID":"44eb39fa-5b20-4c02-b039-f23d7c5220d2","Type":"ContainerStarted","Data":"6d5a5224e568b69a32a45e8e62160cea75b453147a9d56c99f38ac28584df313"} Oct 11 04:05:10 crc kubenswrapper[4967]: I1011 04:05:10.161960 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-zt44b" Oct 11 04:05:10 crc kubenswrapper[4967]: I1011 04:05:10.164459 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-8ccrl" event={"ID":"e9feb54d-dbfa-40ec-ac45-535cca431024","Type":"ContainerStarted","Data":"98d45c2ca67c7b220f2a1d78b3417196e2a0aed830ed3d3277ff64a713755e66"} Oct 11 04:05:10 crc kubenswrapper[4967]: I1011 04:05:10.164573 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-8ccrl" Oct 11 04:05:10 crc kubenswrapper[4967]: I1011 04:05:10.166036 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-6zmv5" event={"ID":"538ebbcb-d51c-472c-a2b4-80236b0c8c08","Type":"ContainerStarted","Data":"d3356b15673a96d1156808958c44a17831e807fad20627c5deb64117c6f6d565"} Oct 11 04:05:10 crc kubenswrapper[4967]: I1011 04:05:10.166574 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-6zmv5" Oct 11 04:05:10 crc kubenswrapper[4967]: I1011 04:05:10.168422 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ironic-operator-controller-manager-67494cb947-sklz6" event={"ID":"674852f1-9729-4a6e-8c90-31741fa69b49","Type":"ContainerStarted","Data":"f1850e7a874d4abb4a453d164165719cca0b59c88cac0c03d2d88058bbf56d3a"} Oct 11 04:05:10 crc kubenswrapper[4967]: I1011 04:05:10.168916 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ironic-operator-controller-manager-67494cb947-sklz6" Oct 11 04:05:10 crc kubenswrapper[4967]: I1011 04:05:10.170886 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-5glgp" event={"ID":"91f45294-2d2f-4b40-a1b4-a841c7748cc8","Type":"ContainerStarted","Data":"f52adbb1b9a15aad7cd50d829f2db60025ce2203577b81c961a8502fdcb103eb"} Oct 11 04:05:10 crc kubenswrapper[4967]: I1011 04:05:10.171054 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-5glgp" Oct 11 04:05:10 crc kubenswrapper[4967]: I1011 04:05:10.174241 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-kkrkc" event={"ID":"c9e29e65-cdd1-408d-bed0-bc283147b283","Type":"ContainerStarted","Data":"2851d9fd1295ab54cb92fd0f0e3ba98b3a3712f87d1bf62768e0fd038a187ac7"} Oct 11 04:05:10 crc kubenswrapper[4967]: I1011 04:05:10.174377 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-kkrkc" Oct 11 04:05:10 crc kubenswrapper[4967]: I1011 04:05:10.176285 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/placement-operator-controller-manager-664664cb68-vng4c" event={"ID":"26b9d5a0-dfbe-4e53-aa03-4fbb4682019b","Type":"ContainerStarted","Data":"60c4544b81b28e7d874fa8dbb3d3cd8ee73658e790231e85ee847e28107bd5af"} Oct 11 04:05:10 crc kubenswrapper[4967]: I1011 04:05:10.183195 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/manila-operator-controller-manager-59578bc799-ls9k6" event={"ID":"20584c10-d8a1-41ca-a0a2-9465fa5190f2","Type":"ContainerStarted","Data":"cf81c3856f2e0b105177b9ce8a87b0339a71e82b5063ae424d6c5aae427cef06"} Oct 11 04:05:10 crc kubenswrapper[4967]: I1011 04:05:10.181600 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-4s7hv" podStartSLOduration=5.157391536 podStartE2EDuration="23.181573813s" podCreationTimestamp="2025-10-11 04:04:47 +0000 UTC" firstStartedPulling="2025-10-11 04:04:49.810346777 +0000 UTC m=+817.773555710" lastFinishedPulling="2025-10-11 04:05:07.834529054 +0000 UTC m=+835.797737987" observedRunningTime="2025-10-11 04:05:09.468611428 +0000 UTC m=+837.431820361" watchObservedRunningTime="2025-10-11 04:05:10.181573813 +0000 UTC m=+838.144782756" Oct 11 04:05:10 crc kubenswrapper[4967]: I1011 04:05:10.183582 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/placement-operator-controller-manager-664664cb68-vng4c" Oct 11 04:05:10 crc kubenswrapper[4967]: I1011 04:05:10.184936 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-9gn9l" event={"ID":"2a2805ea-7d92-42e4-89cb-9b2672f72fc6","Type":"ContainerStarted","Data":"6b91286a1831cf9c9123ab356af2dcd97d42480cea61fe27ea348a05b5d4a4e1"} Oct 11 04:05:10 crc kubenswrapper[4967]: I1011 04:05:10.185516 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-9gn9l" Oct 11 04:05:10 crc kubenswrapper[4967]: E1011 04:05:10.187871 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/heat-operator@sha256:ec11cb8711bd1af22db3c84aa854349ee46191add3db45aecfabb1d8410c04d0\\\"\"" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-ww7nt" podUID="ecd905f3-adb2-47ae-a86d-6fc0b4907a76" Oct 11 04:05:10 crc kubenswrapper[4967]: E1011 04:05:10.188357 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/barbican-operator@sha256:783f711b4cb179819cfcb81167c3591c70671440f4551bbe48b7a8730567f577\\\"\"" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-xz92m" podUID="6f3fbb7d-1207-45a2-8213-c7c71bd703f1" Oct 11 04:05:10 crc kubenswrapper[4967]: E1011 04:05:10.188390 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/nova-operator@sha256:b2e9acf568a48c28cf2aed6012e432eeeb7d5f0eb11878fc91b62bc34cba10cd\\\"\"" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-59px8" podUID="7140d0e1-9937-491b-aa8a-16db830aff11" Oct 11 04:05:10 crc kubenswrapper[4967]: I1011 04:05:10.188396 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/infra-operator-controller-manager-685c45897-5gbwb" podStartSLOduration=6.317007777 podStartE2EDuration="23.188379363s" podCreationTimestamp="2025-10-11 04:04:47 +0000 UTC" firstStartedPulling="2025-10-11 04:04:49.82357424 +0000 UTC m=+817.786783173" lastFinishedPulling="2025-10-11 04:05:06.694945776 +0000 UTC m=+834.658154759" observedRunningTime="2025-10-11 04:05:10.178964792 +0000 UTC m=+838.142173725" watchObservedRunningTime="2025-10-11 04:05:10.188379363 +0000 UTC m=+838.151588296" Oct 11 04:05:10 crc kubenswrapper[4967]: E1011 04:05:10.188443 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"manager\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/openstack-k8s-operators/keystone-operator@sha256:79b43a69884631c635d2164b95a2d4ec68f5cb33f96da14764f1c710880f3997\\\"\"" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-brbrs" podUID="950d3635-f160-45a1-9a4c-749720e662dc" Oct 11 04:05:10 crc kubenswrapper[4967]: I1011 04:05:10.211508 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-kkrkc" podStartSLOduration=6.87534944 podStartE2EDuration="23.211490319s" podCreationTimestamp="2025-10-11 04:04:47 +0000 UTC" firstStartedPulling="2025-10-11 04:04:49.353013089 +0000 UTC m=+817.316222022" lastFinishedPulling="2025-10-11 04:05:05.689153978 +0000 UTC m=+833.652362901" observedRunningTime="2025-10-11 04:05:10.205786515 +0000 UTC m=+838.168995458" watchObservedRunningTime="2025-10-11 04:05:10.211490319 +0000 UTC m=+838.174699262" Oct 11 04:05:10 crc kubenswrapper[4967]: I1011 04:05:10.224377 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-5glgp" podStartSLOduration=6.55049062 podStartE2EDuration="23.224337263s" podCreationTimestamp="2025-10-11 04:04:47 +0000 UTC" firstStartedPulling="2025-10-11 04:04:49.006647 +0000 UTC m=+816.969855933" lastFinishedPulling="2025-10-11 04:05:05.680493633 +0000 UTC m=+833.643702576" observedRunningTime="2025-10-11 04:05:10.221808693 +0000 UTC m=+838.185017626" watchObservedRunningTime="2025-10-11 04:05:10.224337263 +0000 UTC m=+838.187546206" Oct 11 04:05:10 crc kubenswrapper[4967]: I1011 04:05:10.256800 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-8ccrl" podStartSLOduration=7.192118259 podStartE2EDuration="23.256778458s" podCreationTimestamp="2025-10-11 04:04:47 +0000 UTC" firstStartedPulling="2025-10-11 04:04:49.181332245 +0000 UTC m=+817.144541178" lastFinishedPulling="2025-10-11 04:05:05.245992444 +0000 UTC m=+833.209201377" observedRunningTime="2025-10-11 04:05:10.24666136 +0000 UTC m=+838.209870293" watchObservedRunningTime="2025-10-11 04:05:10.256778458 +0000 UTC m=+838.219987391" Oct 11 04:05:10 crc kubenswrapper[4967]: I1011 04:05:10.268639 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-zt44b" podStartSLOduration=6.586667664 podStartE2EDuration="23.268624739s" podCreationTimestamp="2025-10-11 04:04:47 +0000 UTC" firstStartedPulling="2025-10-11 04:04:49.00708317 +0000 UTC m=+816.970292103" lastFinishedPulling="2025-10-11 04:05:05.689040205 +0000 UTC m=+833.652249178" observedRunningTime="2025-10-11 04:05:10.263361854 +0000 UTC m=+838.226570797" watchObservedRunningTime="2025-10-11 04:05:10.268624739 +0000 UTC m=+838.231833672" Oct 11 04:05:10 crc kubenswrapper[4967]: I1011 04:05:10.297619 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/ironic-operator-controller-manager-67494cb947-sklz6" podStartSLOduration=6.62294281 podStartE2EDuration="23.297599672s" podCreationTimestamp="2025-10-11 04:04:47 +0000 UTC" firstStartedPulling="2025-10-11 04:04:49.006607559 +0000 UTC m=+816.969816492" lastFinishedPulling="2025-10-11 04:05:05.681264421 +0000 UTC m=+833.644473354" observedRunningTime="2025-10-11 04:05:10.281750758 +0000 UTC m=+838.244959701" watchObservedRunningTime="2025-10-11 04:05:10.297599672 +0000 UTC m=+838.260808605" Oct 11 04:05:10 crc kubenswrapper[4967]: I1011 04:05:10.298374 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-6zmv5" podStartSLOduration=6.874362548 podStartE2EDuration="23.298365871s" podCreationTimestamp="2025-10-11 04:04:47 +0000 UTC" firstStartedPulling="2025-10-11 04:04:49.739494665 +0000 UTC m=+817.702703598" lastFinishedPulling="2025-10-11 04:05:06.163497988 +0000 UTC m=+834.126706921" observedRunningTime="2025-10-11 04:05:10.294855147 +0000 UTC m=+838.258064080" watchObservedRunningTime="2025-10-11 04:05:10.298365871 +0000 UTC m=+838.261574804" Oct 11 04:05:10 crc kubenswrapper[4967]: I1011 04:05:10.309175 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-9gn9l" podStartSLOduration=6.061544376 podStartE2EDuration="23.309159806s" podCreationTimestamp="2025-10-11 04:04:47 +0000 UTC" firstStartedPulling="2025-10-11 04:04:48.442048419 +0000 UTC m=+816.405257352" lastFinishedPulling="2025-10-11 04:05:05.689663849 +0000 UTC m=+833.652872782" observedRunningTime="2025-10-11 04:05:10.306891322 +0000 UTC m=+838.270100265" watchObservedRunningTime="2025-10-11 04:05:10.309159806 +0000 UTC m=+838.272368739" Oct 11 04:05:10 crc kubenswrapper[4967]: I1011 04:05:10.364032 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/placement-operator-controller-manager-664664cb68-vng4c" podStartSLOduration=5.265992442 podStartE2EDuration="23.364016481s" podCreationTimestamp="2025-10-11 04:04:47 +0000 UTC" firstStartedPulling="2025-10-11 04:04:49.803859085 +0000 UTC m=+817.767068018" lastFinishedPulling="2025-10-11 04:05:07.901883124 +0000 UTC m=+835.865092057" observedRunningTime="2025-10-11 04:05:10.360364294 +0000 UTC m=+838.323573227" watchObservedRunningTime="2025-10-11 04:05:10.364016481 +0000 UTC m=+838.327225404" Oct 11 04:05:11 crc kubenswrapper[4967]: I1011 04:05:11.197383 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-fgz82" event={"ID":"e82102ad-391b-4b29-ae47-dae18dba7743","Type":"ContainerStarted","Data":"103a0d8fcd1681db6ebfa644f6e7eaa0bb736a09e8e836995cd5c373d267240c"} Oct 11 04:05:11 crc kubenswrapper[4967]: I1011 04:05:11.216616 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-fgz82" podStartSLOduration=3.468792744 podStartE2EDuration="24.216599222s" podCreationTimestamp="2025-10-11 04:04:47 +0000 UTC" firstStartedPulling="2025-10-11 04:04:49.775391662 +0000 UTC m=+817.738600595" lastFinishedPulling="2025-10-11 04:05:10.52319814 +0000 UTC m=+838.486407073" observedRunningTime="2025-10-11 04:05:11.209905484 +0000 UTC m=+839.173114457" watchObservedRunningTime="2025-10-11 04:05:11.216599222 +0000 UTC m=+839.179808155" Oct 11 04:05:12 crc kubenswrapper[4967]: I1011 04:05:12.210957 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-65knj" event={"ID":"1094a8be-16bb-4986-b4a4-465d41880800","Type":"ContainerStarted","Data":"08397172fadaaf51d9962a3d6a14b1f8027a1c47c59bd4eac9c9a26cd450f7fb"} Oct 11 04:05:17 crc kubenswrapper[4967]: I1011 04:05:17.809354 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/designate-operator-controller-manager-687df44cdb-9gn9l" Oct 11 04:05:17 crc kubenswrapper[4967]: I1011 04:05:17.837621 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-65knj" podStartSLOduration=9.06398859 podStartE2EDuration="30.837598198s" podCreationTimestamp="2025-10-11 04:04:47 +0000 UTC" firstStartedPulling="2025-10-11 04:04:49.843734756 +0000 UTC m=+817.806943689" lastFinishedPulling="2025-10-11 04:05:11.617344354 +0000 UTC m=+839.580553297" observedRunningTime="2025-10-11 04:05:12.242975237 +0000 UTC m=+840.206184200" watchObservedRunningTime="2025-10-11 04:05:17.837598198 +0000 UTC m=+845.800807122" Oct 11 04:05:18 crc kubenswrapper[4967]: I1011 04:05:18.005690 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ironic-operator-controller-manager-67494cb947-sklz6" Oct 11 04:05:18 crc kubenswrapper[4967]: I1011 04:05:18.079970 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/cinder-operator-controller-manager-59cdc64769-5glgp" Oct 11 04:05:18 crc kubenswrapper[4967]: I1011 04:05:18.158227 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/glance-operator-controller-manager-7bb46cd7d-zt44b" Oct 11 04:05:18 crc kubenswrapper[4967]: I1011 04:05:18.228648 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/horizon-operator-controller-manager-6d74794d9b-8ccrl" Oct 11 04:05:18 crc kubenswrapper[4967]: I1011 04:05:18.229144 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/manila-operator-controller-manager-59578bc799-ls9k6" Oct 11 04:05:18 crc kubenswrapper[4967]: I1011 04:05:18.304616 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/neutron-operator-controller-manager-797d478b46-6zmv5" Oct 11 04:05:18 crc kubenswrapper[4967]: I1011 04:05:18.383592 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/octavia-operator-controller-manager-6d7c7ddf95-7clg2" Oct 11 04:05:18 crc kubenswrapper[4967]: I1011 04:05:18.436896 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/ovn-operator-controller-manager-869cc7797f-kkrkc" Oct 11 04:05:18 crc kubenswrapper[4967]: I1011 04:05:18.492554 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/placement-operator-controller-manager-664664cb68-vng4c" Oct 11 04:05:18 crc kubenswrapper[4967]: I1011 04:05:18.561678 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/infra-operator-controller-manager-685c45897-5gbwb" Oct 11 04:05:18 crc kubenswrapper[4967]: I1011 04:05:18.597630 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-65knj" Oct 11 04:05:18 crc kubenswrapper[4967]: I1011 04:05:18.600115 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/swift-operator-controller-manager-5f4d5dfdc6-65knj" Oct 11 04:05:18 crc kubenswrapper[4967]: I1011 04:05:18.747729 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-fgz82" Oct 11 04:05:18 crc kubenswrapper[4967]: I1011 04:05:18.749567 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/telemetry-operator-controller-manager-578874c84d-fgz82" Oct 11 04:05:18 crc kubenswrapper[4967]: I1011 04:05:18.860289 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/test-operator-controller-manager-ffcdd6c94-4s7hv" Oct 11 04:05:18 crc kubenswrapper[4967]: I1011 04:05:18.864787 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/watcher-operator-controller-manager-646675d848-zdpbn" Oct 11 04:05:19 crc kubenswrapper[4967]: I1011 04:05:19.020117 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns" Oct 11 04:05:22 crc kubenswrapper[4967]: I1011 04:05:22.304706 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-brbrs" event={"ID":"950d3635-f160-45a1-9a4c-749720e662dc","Type":"ContainerStarted","Data":"0f92919342cbe91f90b14e0a05a970e661ad4b7961fd267fa183588e3be947ec"} Oct 11 04:05:22 crc kubenswrapper[4967]: I1011 04:05:22.305874 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-brbrs" Oct 11 04:05:22 crc kubenswrapper[4967]: I1011 04:05:22.325438 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-brbrs" podStartSLOduration=3.264129293 podStartE2EDuration="35.325412865s" podCreationTimestamp="2025-10-11 04:04:47 +0000 UTC" firstStartedPulling="2025-10-11 04:04:49.343280839 +0000 UTC m=+817.306489772" lastFinishedPulling="2025-10-11 04:05:21.404564411 +0000 UTC m=+849.367773344" observedRunningTime="2025-10-11 04:05:22.321547163 +0000 UTC m=+850.284756086" watchObservedRunningTime="2025-10-11 04:05:22.325412865 +0000 UTC m=+850.288621838" Oct 11 04:05:24 crc kubenswrapper[4967]: I1011 04:05:24.321520 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-ww7nt" event={"ID":"ecd905f3-adb2-47ae-a86d-6fc0b4907a76","Type":"ContainerStarted","Data":"6d64025f43be276fc4eb6117c9f6ff27e72ade625813b33e5bd523106b4028b7"} Oct 11 04:05:24 crc kubenswrapper[4967]: I1011 04:05:24.322373 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-ww7nt" Oct 11 04:05:24 crc kubenswrapper[4967]: I1011 04:05:24.324827 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-fjr8k" event={"ID":"ef8ae8f4-ad95-4f1f-bc8c-170fe5058dc9","Type":"ContainerStarted","Data":"0b86aa806e7a82f0d47e72766e6845ced57c3260c12155d594f99885025a6e05"} Oct 11 04:05:24 crc kubenswrapper[4967]: I1011 04:05:24.325119 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-fjr8k" Oct 11 04:05:24 crc kubenswrapper[4967]: I1011 04:05:24.326917 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-59px8" event={"ID":"7140d0e1-9937-491b-aa8a-16db830aff11","Type":"ContainerStarted","Data":"63e45c018907b38c1ab2bc437eb8c0ca8ac72553a32ea0c14bcf3128601985ce"} Oct 11 04:05:24 crc kubenswrapper[4967]: I1011 04:05:24.327151 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-59px8" Oct 11 04:05:24 crc kubenswrapper[4967]: I1011 04:05:24.341994 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-ww7nt" podStartSLOduration=3.062222295 podStartE2EDuration="37.341978339s" podCreationTimestamp="2025-10-11 04:04:47 +0000 UTC" firstStartedPulling="2025-10-11 04:04:49.035921542 +0000 UTC m=+816.999130465" lastFinishedPulling="2025-10-11 04:05:23.315677576 +0000 UTC m=+851.278886509" observedRunningTime="2025-10-11 04:05:24.338925158 +0000 UTC m=+852.302134091" watchObservedRunningTime="2025-10-11 04:05:24.341978339 +0000 UTC m=+852.305187272" Oct 11 04:05:24 crc kubenswrapper[4967]: I1011 04:05:24.357033 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-59px8" podStartSLOduration=3.498110597 podStartE2EDuration="37.357011545s" podCreationTimestamp="2025-10-11 04:04:47 +0000 UTC" firstStartedPulling="2025-10-11 04:04:49.38101659 +0000 UTC m=+817.344225523" lastFinishedPulling="2025-10-11 04:05:23.239917538 +0000 UTC m=+851.203126471" observedRunningTime="2025-10-11 04:05:24.356577165 +0000 UTC m=+852.319786138" watchObservedRunningTime="2025-10-11 04:05:24.357011545 +0000 UTC m=+852.320220478" Oct 11 04:05:24 crc kubenswrapper[4967]: I1011 04:05:24.382149 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-fjr8k" podStartSLOduration=3.3606554109999998 podStartE2EDuration="37.382124738s" podCreationTimestamp="2025-10-11 04:04:47 +0000 UTC" firstStartedPulling="2025-10-11 04:04:49.218183705 +0000 UTC m=+817.181392638" lastFinishedPulling="2025-10-11 04:05:23.239653012 +0000 UTC m=+851.202861965" observedRunningTime="2025-10-11 04:05:24.375612114 +0000 UTC m=+852.338821077" watchObservedRunningTime="2025-10-11 04:05:24.382124738 +0000 UTC m=+852.345333711" Oct 11 04:05:25 crc kubenswrapper[4967]: I1011 04:05:25.341638 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-xz92m" event={"ID":"6f3fbb7d-1207-45a2-8213-c7c71bd703f1","Type":"ContainerStarted","Data":"7bff5bbd15e78ee0efb9297587de47a2ac22d4a50a94fcefe60f9d68b8fad10f"} Oct 11 04:05:25 crc kubenswrapper[4967]: I1011 04:05:25.373765 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-xz92m" podStartSLOduration=2.935026931 podStartE2EDuration="38.373738471s" podCreationTimestamp="2025-10-11 04:04:47 +0000 UTC" firstStartedPulling="2025-10-11 04:04:48.823470335 +0000 UTC m=+816.786679268" lastFinishedPulling="2025-10-11 04:05:24.262181875 +0000 UTC m=+852.225390808" observedRunningTime="2025-10-11 04:05:25.365346933 +0000 UTC m=+853.328555896" watchObservedRunningTime="2025-10-11 04:05:25.373738471 +0000 UTC m=+853.336947444" Oct 11 04:05:27 crc kubenswrapper[4967]: I1011 04:05:27.768463 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-xz92m" Oct 11 04:05:28 crc kubenswrapper[4967]: I1011 04:05:28.165195 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/keystone-operator-controller-manager-ddb98f99b-brbrs" Oct 11 04:05:28 crc kubenswrapper[4967]: I1011 04:05:28.254228 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/mariadb-operator-controller-manager-5777b4f897-fjr8k" Oct 11 04:05:28 crc kubenswrapper[4967]: I1011 04:05:28.326939 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/nova-operator-controller-manager-57bb74c7bf-59px8" Oct 11 04:05:37 crc kubenswrapper[4967]: I1011 04:05:37.769968 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/barbican-operator-controller-manager-64f84fcdbb-xz92m" Oct 11 04:05:37 crc kubenswrapper[4967]: I1011 04:05:37.912009 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack-operators/heat-operator-controller-manager-6d9967f8dd-ww7nt" Oct 11 04:05:44 crc kubenswrapper[4967]: I1011 04:05:44.089326 4967 patch_prober.go:28] interesting pod/machine-config-daemon-sgrv2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:05:44 crc kubenswrapper[4967]: I1011 04:05:44.089971 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.505471 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lxnrr"] Oct 11 04:05:56 crc kubenswrapper[4967]: E1011 04:05:56.506367 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7de04fa6-4dd6-4254-bde0-b83b920be0fa" containerName="registry-server" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.506384 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="7de04fa6-4dd6-4254-bde0-b83b920be0fa" containerName="registry-server" Oct 11 04:05:56 crc kubenswrapper[4967]: E1011 04:05:56.506427 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7de04fa6-4dd6-4254-bde0-b83b920be0fa" containerName="extract-content" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.506457 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="7de04fa6-4dd6-4254-bde0-b83b920be0fa" containerName="extract-content" Oct 11 04:05:56 crc kubenswrapper[4967]: E1011 04:05:56.506494 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8487ea91-dfea-423a-8c19-358a235b9eff" containerName="extract-content" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.506503 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="8487ea91-dfea-423a-8c19-358a235b9eff" containerName="extract-content" Oct 11 04:05:56 crc kubenswrapper[4967]: E1011 04:05:56.506519 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7de04fa6-4dd6-4254-bde0-b83b920be0fa" containerName="extract-utilities" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.506528 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="7de04fa6-4dd6-4254-bde0-b83b920be0fa" containerName="extract-utilities" Oct 11 04:05:56 crc kubenswrapper[4967]: E1011 04:05:56.506549 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8487ea91-dfea-423a-8c19-358a235b9eff" containerName="registry-server" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.506558 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="8487ea91-dfea-423a-8c19-358a235b9eff" containerName="registry-server" Oct 11 04:05:56 crc kubenswrapper[4967]: E1011 04:05:56.506579 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8487ea91-dfea-423a-8c19-358a235b9eff" containerName="extract-utilities" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.506587 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="8487ea91-dfea-423a-8c19-358a235b9eff" containerName="extract-utilities" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.506764 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="8487ea91-dfea-423a-8c19-358a235b9eff" containerName="registry-server" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.506790 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="7de04fa6-4dd6-4254-bde0-b83b920be0fa" containerName="registry-server" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.518022 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-lxnrr" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.523194 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"dnsmasq-dns-dockercfg-ghv8w" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.523827 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"kube-root-ca.crt" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.523955 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.525642 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lxnrr"] Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.528162 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openshift-service-ca.crt" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.562417 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-8wsqp"] Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.563496 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-8wsqp" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.565621 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-svc" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.580941 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-8wsqp"] Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.648358 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f980d5dd-67f3-4714-a626-d521efde4916-config\") pod \"dnsmasq-dns-675f4bcbfc-lxnrr\" (UID: \"f980d5dd-67f3-4714-a626-d521efde4916\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lxnrr" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.648559 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqpp9\" (UniqueName: \"kubernetes.io/projected/f980d5dd-67f3-4714-a626-d521efde4916-kube-api-access-gqpp9\") pod \"dnsmasq-dns-675f4bcbfc-lxnrr\" (UID: \"f980d5dd-67f3-4714-a626-d521efde4916\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lxnrr" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.749577 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4s4ft\" (UniqueName: \"kubernetes.io/projected/854c2370-c43a-4c50-99ab-5d935456c2ae-kube-api-access-4s4ft\") pod \"dnsmasq-dns-78dd6ddcc-8wsqp\" (UID: \"854c2370-c43a-4c50-99ab-5d935456c2ae\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8wsqp" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.749645 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqpp9\" (UniqueName: \"kubernetes.io/projected/f980d5dd-67f3-4714-a626-d521efde4916-kube-api-access-gqpp9\") pod \"dnsmasq-dns-675f4bcbfc-lxnrr\" (UID: \"f980d5dd-67f3-4714-a626-d521efde4916\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lxnrr" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.749672 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/854c2370-c43a-4c50-99ab-5d935456c2ae-config\") pod \"dnsmasq-dns-78dd6ddcc-8wsqp\" (UID: \"854c2370-c43a-4c50-99ab-5d935456c2ae\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8wsqp" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.749693 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/854c2370-c43a-4c50-99ab-5d935456c2ae-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-8wsqp\" (UID: \"854c2370-c43a-4c50-99ab-5d935456c2ae\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8wsqp" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.750039 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f980d5dd-67f3-4714-a626-d521efde4916-config\") pod \"dnsmasq-dns-675f4bcbfc-lxnrr\" (UID: \"f980d5dd-67f3-4714-a626-d521efde4916\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lxnrr" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.750930 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f980d5dd-67f3-4714-a626-d521efde4916-config\") pod \"dnsmasq-dns-675f4bcbfc-lxnrr\" (UID: \"f980d5dd-67f3-4714-a626-d521efde4916\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lxnrr" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.772911 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqpp9\" (UniqueName: \"kubernetes.io/projected/f980d5dd-67f3-4714-a626-d521efde4916-kube-api-access-gqpp9\") pod \"dnsmasq-dns-675f4bcbfc-lxnrr\" (UID: \"f980d5dd-67f3-4714-a626-d521efde4916\") " pod="openstack/dnsmasq-dns-675f4bcbfc-lxnrr" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.838163 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-lxnrr" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.851531 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4s4ft\" (UniqueName: \"kubernetes.io/projected/854c2370-c43a-4c50-99ab-5d935456c2ae-kube-api-access-4s4ft\") pod \"dnsmasq-dns-78dd6ddcc-8wsqp\" (UID: \"854c2370-c43a-4c50-99ab-5d935456c2ae\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8wsqp" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.851804 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/854c2370-c43a-4c50-99ab-5d935456c2ae-config\") pod \"dnsmasq-dns-78dd6ddcc-8wsqp\" (UID: \"854c2370-c43a-4c50-99ab-5d935456c2ae\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8wsqp" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.851855 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/854c2370-c43a-4c50-99ab-5d935456c2ae-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-8wsqp\" (UID: \"854c2370-c43a-4c50-99ab-5d935456c2ae\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8wsqp" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.852658 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/854c2370-c43a-4c50-99ab-5d935456c2ae-dns-svc\") pod \"dnsmasq-dns-78dd6ddcc-8wsqp\" (UID: \"854c2370-c43a-4c50-99ab-5d935456c2ae\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8wsqp" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.852935 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/854c2370-c43a-4c50-99ab-5d935456c2ae-config\") pod \"dnsmasq-dns-78dd6ddcc-8wsqp\" (UID: \"854c2370-c43a-4c50-99ab-5d935456c2ae\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8wsqp" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.875278 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4s4ft\" (UniqueName: \"kubernetes.io/projected/854c2370-c43a-4c50-99ab-5d935456c2ae-kube-api-access-4s4ft\") pod \"dnsmasq-dns-78dd6ddcc-8wsqp\" (UID: \"854c2370-c43a-4c50-99ab-5d935456c2ae\") " pod="openstack/dnsmasq-dns-78dd6ddcc-8wsqp" Oct 11 04:05:56 crc kubenswrapper[4967]: I1011 04:05:56.880302 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-8wsqp" Oct 11 04:05:57 crc kubenswrapper[4967]: I1011 04:05:57.312787 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lxnrr"] Oct 11 04:05:57 crc kubenswrapper[4967]: I1011 04:05:57.427501 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-8wsqp"] Oct 11 04:05:57 crc kubenswrapper[4967]: I1011 04:05:57.629162 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-8wsqp" event={"ID":"854c2370-c43a-4c50-99ab-5d935456c2ae","Type":"ContainerStarted","Data":"a6bc70e9455375f1ba08ce5d78aacef5cee9a5ce589d9945d3399b44feeb1472"} Oct 11 04:05:57 crc kubenswrapper[4967]: I1011 04:05:57.630447 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-lxnrr" event={"ID":"f980d5dd-67f3-4714-a626-d521efde4916","Type":"ContainerStarted","Data":"846dde0fc4421a635bdd8a534ba89c85cf76194c7b4475a1fe486b0aa603fb23"} Oct 11 04:05:59 crc kubenswrapper[4967]: I1011 04:05:59.424176 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lxnrr"] Oct 11 04:05:59 crc kubenswrapper[4967]: I1011 04:05:59.443949 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-vc6g6"] Oct 11 04:05:59 crc kubenswrapper[4967]: I1011 04:05:59.445035 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-vc6g6" Oct 11 04:05:59 crc kubenswrapper[4967]: I1011 04:05:59.514346 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-vc6g6"] Oct 11 04:05:59 crc kubenswrapper[4967]: I1011 04:05:59.631816 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9b792fe-ab24-4587-a1f6-2ad194c9eb83-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-vc6g6\" (UID: \"b9b792fe-ab24-4587-a1f6-2ad194c9eb83\") " pod="openstack/dnsmasq-dns-5ccc8479f9-vc6g6" Oct 11 04:05:59 crc kubenswrapper[4967]: I1011 04:05:59.631863 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9b792fe-ab24-4587-a1f6-2ad194c9eb83-config\") pod \"dnsmasq-dns-5ccc8479f9-vc6g6\" (UID: \"b9b792fe-ab24-4587-a1f6-2ad194c9eb83\") " pod="openstack/dnsmasq-dns-5ccc8479f9-vc6g6" Oct 11 04:05:59 crc kubenswrapper[4967]: I1011 04:05:59.631887 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jdbwh\" (UniqueName: \"kubernetes.io/projected/b9b792fe-ab24-4587-a1f6-2ad194c9eb83-kube-api-access-jdbwh\") pod \"dnsmasq-dns-5ccc8479f9-vc6g6\" (UID: \"b9b792fe-ab24-4587-a1f6-2ad194c9eb83\") " pod="openstack/dnsmasq-dns-5ccc8479f9-vc6g6" Oct 11 04:05:59 crc kubenswrapper[4967]: I1011 04:05:59.711471 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-8wsqp"] Oct 11 04:05:59 crc kubenswrapper[4967]: I1011 04:05:59.733238 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9b792fe-ab24-4587-a1f6-2ad194c9eb83-config\") pod \"dnsmasq-dns-5ccc8479f9-vc6g6\" (UID: \"b9b792fe-ab24-4587-a1f6-2ad194c9eb83\") " pod="openstack/dnsmasq-dns-5ccc8479f9-vc6g6" Oct 11 04:05:59 crc kubenswrapper[4967]: I1011 04:05:59.733282 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jdbwh\" (UniqueName: \"kubernetes.io/projected/b9b792fe-ab24-4587-a1f6-2ad194c9eb83-kube-api-access-jdbwh\") pod \"dnsmasq-dns-5ccc8479f9-vc6g6\" (UID: \"b9b792fe-ab24-4587-a1f6-2ad194c9eb83\") " pod="openstack/dnsmasq-dns-5ccc8479f9-vc6g6" Oct 11 04:05:59 crc kubenswrapper[4967]: I1011 04:05:59.733377 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9b792fe-ab24-4587-a1f6-2ad194c9eb83-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-vc6g6\" (UID: \"b9b792fe-ab24-4587-a1f6-2ad194c9eb83\") " pod="openstack/dnsmasq-dns-5ccc8479f9-vc6g6" Oct 11 04:05:59 crc kubenswrapper[4967]: I1011 04:05:59.734855 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9b792fe-ab24-4587-a1f6-2ad194c9eb83-dns-svc\") pod \"dnsmasq-dns-5ccc8479f9-vc6g6\" (UID: \"b9b792fe-ab24-4587-a1f6-2ad194c9eb83\") " pod="openstack/dnsmasq-dns-5ccc8479f9-vc6g6" Oct 11 04:05:59 crc kubenswrapper[4967]: I1011 04:05:59.734951 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9b792fe-ab24-4587-a1f6-2ad194c9eb83-config\") pod \"dnsmasq-dns-5ccc8479f9-vc6g6\" (UID: \"b9b792fe-ab24-4587-a1f6-2ad194c9eb83\") " pod="openstack/dnsmasq-dns-5ccc8479f9-vc6g6" Oct 11 04:05:59 crc kubenswrapper[4967]: I1011 04:05:59.740558 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5vbgv"] Oct 11 04:05:59 crc kubenswrapper[4967]: I1011 04:05:59.744183 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-5vbgv" Oct 11 04:05:59 crc kubenswrapper[4967]: I1011 04:05:59.757991 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jdbwh\" (UniqueName: \"kubernetes.io/projected/b9b792fe-ab24-4587-a1f6-2ad194c9eb83-kube-api-access-jdbwh\") pod \"dnsmasq-dns-5ccc8479f9-vc6g6\" (UID: \"b9b792fe-ab24-4587-a1f6-2ad194c9eb83\") " pod="openstack/dnsmasq-dns-5ccc8479f9-vc6g6" Oct 11 04:05:59 crc kubenswrapper[4967]: I1011 04:05:59.759356 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5vbgv"] Oct 11 04:05:59 crc kubenswrapper[4967]: I1011 04:05:59.785288 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-vc6g6" Oct 11 04:05:59 crc kubenswrapper[4967]: I1011 04:05:59.843886 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/803e1df8-1759-49e3-83c8-abb1abf6fff4-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-5vbgv\" (UID: \"803e1df8-1759-49e3-83c8-abb1abf6fff4\") " pod="openstack/dnsmasq-dns-57d769cc4f-5vbgv" Oct 11 04:05:59 crc kubenswrapper[4967]: I1011 04:05:59.844594 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f66tt\" (UniqueName: \"kubernetes.io/projected/803e1df8-1759-49e3-83c8-abb1abf6fff4-kube-api-access-f66tt\") pod \"dnsmasq-dns-57d769cc4f-5vbgv\" (UID: \"803e1df8-1759-49e3-83c8-abb1abf6fff4\") " pod="openstack/dnsmasq-dns-57d769cc4f-5vbgv" Oct 11 04:05:59 crc kubenswrapper[4967]: I1011 04:05:59.844640 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/803e1df8-1759-49e3-83c8-abb1abf6fff4-config\") pod \"dnsmasq-dns-57d769cc4f-5vbgv\" (UID: \"803e1df8-1759-49e3-83c8-abb1abf6fff4\") " pod="openstack/dnsmasq-dns-57d769cc4f-5vbgv" Oct 11 04:05:59 crc kubenswrapper[4967]: I1011 04:05:59.948395 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f66tt\" (UniqueName: \"kubernetes.io/projected/803e1df8-1759-49e3-83c8-abb1abf6fff4-kube-api-access-f66tt\") pod \"dnsmasq-dns-57d769cc4f-5vbgv\" (UID: \"803e1df8-1759-49e3-83c8-abb1abf6fff4\") " pod="openstack/dnsmasq-dns-57d769cc4f-5vbgv" Oct 11 04:05:59 crc kubenswrapper[4967]: I1011 04:05:59.948472 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/803e1df8-1759-49e3-83c8-abb1abf6fff4-config\") pod \"dnsmasq-dns-57d769cc4f-5vbgv\" (UID: \"803e1df8-1759-49e3-83c8-abb1abf6fff4\") " pod="openstack/dnsmasq-dns-57d769cc4f-5vbgv" Oct 11 04:05:59 crc kubenswrapper[4967]: I1011 04:05:59.948531 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/803e1df8-1759-49e3-83c8-abb1abf6fff4-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-5vbgv\" (UID: \"803e1df8-1759-49e3-83c8-abb1abf6fff4\") " pod="openstack/dnsmasq-dns-57d769cc4f-5vbgv" Oct 11 04:05:59 crc kubenswrapper[4967]: I1011 04:05:59.949858 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/803e1df8-1759-49e3-83c8-abb1abf6fff4-config\") pod \"dnsmasq-dns-57d769cc4f-5vbgv\" (UID: \"803e1df8-1759-49e3-83c8-abb1abf6fff4\") " pod="openstack/dnsmasq-dns-57d769cc4f-5vbgv" Oct 11 04:05:59 crc kubenswrapper[4967]: I1011 04:05:59.949883 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/803e1df8-1759-49e3-83c8-abb1abf6fff4-dns-svc\") pod \"dnsmasq-dns-57d769cc4f-5vbgv\" (UID: \"803e1df8-1759-49e3-83c8-abb1abf6fff4\") " pod="openstack/dnsmasq-dns-57d769cc4f-5vbgv" Oct 11 04:05:59 crc kubenswrapper[4967]: I1011 04:05:59.969166 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f66tt\" (UniqueName: \"kubernetes.io/projected/803e1df8-1759-49e3-83c8-abb1abf6fff4-kube-api-access-f66tt\") pod \"dnsmasq-dns-57d769cc4f-5vbgv\" (UID: \"803e1df8-1759-49e3-83c8-abb1abf6fff4\") " pod="openstack/dnsmasq-dns-57d769cc4f-5vbgv" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.094887 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-5vbgv" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.246433 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-vc6g6"] Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.586664 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.587960 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.592653 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-wmdzf" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.592970 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.593144 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.593292 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.596281 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.596420 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.597642 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.602391 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.759208 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0c315653-59d1-429f-be9c-c1ac826cff49-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.759248 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0c315653-59d1-429f-be9c-c1ac826cff49-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.759314 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.759335 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0c315653-59d1-429f-be9c-c1ac826cff49-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.759362 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0c315653-59d1-429f-be9c-c1ac826cff49-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.759379 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0c315653-59d1-429f-be9c-c1ac826cff49-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.759407 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0c315653-59d1-429f-be9c-c1ac826cff49-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.759450 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c315653-59d1-429f-be9c-c1ac826cff49-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.759474 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2x4x\" (UniqueName: \"kubernetes.io/projected/0c315653-59d1-429f-be9c-c1ac826cff49-kube-api-access-h2x4x\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.759490 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0c315653-59d1-429f-be9c-c1ac826cff49-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.759508 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0c315653-59d1-429f-be9c-c1ac826cff49-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.853501 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.854890 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.857621 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.858043 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.858319 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-zdwth" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.858376 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.858543 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.858671 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.860493 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.861243 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0c315653-59d1-429f-be9c-c1ac826cff49-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.861296 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0c315653-59d1-429f-be9c-c1ac826cff49-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.861312 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0c315653-59d1-429f-be9c-c1ac826cff49-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.861349 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.861367 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0c315653-59d1-429f-be9c-c1ac826cff49-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.861393 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0c315653-59d1-429f-be9c-c1ac826cff49-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.861410 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0c315653-59d1-429f-be9c-c1ac826cff49-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.861426 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0c315653-59d1-429f-be9c-c1ac826cff49-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.861449 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c315653-59d1-429f-be9c-c1ac826cff49-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.861480 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2x4x\" (UniqueName: \"kubernetes.io/projected/0c315653-59d1-429f-be9c-c1ac826cff49-kube-api-access-h2x4x\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.861499 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0c315653-59d1-429f-be9c-c1ac826cff49-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.862306 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0c315653-59d1-429f-be9c-c1ac826cff49-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.866381 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0c315653-59d1-429f-be9c-c1ac826cff49-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.866398 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0c315653-59d1-429f-be9c-c1ac826cff49-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.866648 4967 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.867764 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c315653-59d1-429f-be9c-c1ac826cff49-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.867121 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0c315653-59d1-429f-be9c-c1ac826cff49-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.869583 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0c315653-59d1-429f-be9c-c1ac826cff49-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.870399 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0c315653-59d1-429f-be9c-c1ac826cff49-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.871455 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0c315653-59d1-429f-be9c-c1ac826cff49-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.872655 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0c315653-59d1-429f-be9c-c1ac826cff49-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.877932 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.888277 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2x4x\" (UniqueName: \"kubernetes.io/projected/0c315653-59d1-429f-be9c-c1ac826cff49-kube-api-access-h2x4x\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.889864 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.921222 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.962914 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g6lk6\" (UniqueName: \"kubernetes.io/projected/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-kube-api-access-g6lk6\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.962954 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.962981 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.963023 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.963065 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.963103 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.963152 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-config-data\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.963167 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.963232 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.963250 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:00 crc kubenswrapper[4967]: I1011 04:06:00.963285 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:01 crc kubenswrapper[4967]: I1011 04:06:01.064710 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:01 crc kubenswrapper[4967]: I1011 04:06:01.064782 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:01 crc kubenswrapper[4967]: I1011 04:06:01.064818 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:01 crc kubenswrapper[4967]: I1011 04:06:01.064846 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:01 crc kubenswrapper[4967]: I1011 04:06:01.064876 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-config-data\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:01 crc kubenswrapper[4967]: I1011 04:06:01.064898 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:01 crc kubenswrapper[4967]: I1011 04:06:01.064928 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:01 crc kubenswrapper[4967]: I1011 04:06:01.064955 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:01 crc kubenswrapper[4967]: I1011 04:06:01.064982 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:01 crc kubenswrapper[4967]: I1011 04:06:01.065032 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g6lk6\" (UniqueName: \"kubernetes.io/projected/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-kube-api-access-g6lk6\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:01 crc kubenswrapper[4967]: I1011 04:06:01.065054 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:01 crc kubenswrapper[4967]: I1011 04:06:01.065790 4967 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-server-0" Oct 11 04:06:01 crc kubenswrapper[4967]: I1011 04:06:01.066210 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:01 crc kubenswrapper[4967]: I1011 04:06:01.066339 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:01 crc kubenswrapper[4967]: I1011 04:06:01.067978 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:01 crc kubenswrapper[4967]: I1011 04:06:01.070164 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-server-conf\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:01 crc kubenswrapper[4967]: I1011 04:06:01.071545 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-config-data\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:01 crc kubenswrapper[4967]: I1011 04:06:01.071967 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:01 crc kubenswrapper[4967]: I1011 04:06:01.071999 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:01 crc kubenswrapper[4967]: I1011 04:06:01.072666 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:01 crc kubenswrapper[4967]: I1011 04:06:01.077611 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-pod-info\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:01 crc kubenswrapper[4967]: I1011 04:06:01.086213 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:01 crc kubenswrapper[4967]: I1011 04:06:01.088542 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g6lk6\" (UniqueName: \"kubernetes.io/projected/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-kube-api-access-g6lk6\") pod \"rabbitmq-server-0\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " pod="openstack/rabbitmq-server-0" Oct 11 04:06:01 crc kubenswrapper[4967]: I1011 04:06:01.243837 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.587232 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-galera-0"] Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.588665 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.591736 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"osp-secret" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.592020 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config-data" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.592291 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-dockercfg-5kdkt" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.593249 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-svc" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.593409 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-scripts" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.596583 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"combined-ca-bundle" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.604258 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.694489 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddc18aa7-8599-40c5-a019-972c13162f83-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ddc18aa7-8599-40c5-a019-972c13162f83\") " pod="openstack/openstack-galera-0" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.694553 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/ddc18aa7-8599-40c5-a019-972c13162f83-secrets\") pod \"openstack-galera-0\" (UID: \"ddc18aa7-8599-40c5-a019-972c13162f83\") " pod="openstack/openstack-galera-0" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.694579 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"ddc18aa7-8599-40c5-a019-972c13162f83\") " pod="openstack/openstack-galera-0" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.694669 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ddc18aa7-8599-40c5-a019-972c13162f83-kolla-config\") pod \"openstack-galera-0\" (UID: \"ddc18aa7-8599-40c5-a019-972c13162f83\") " pod="openstack/openstack-galera-0" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.694723 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ddc18aa7-8599-40c5-a019-972c13162f83-config-data-default\") pod \"openstack-galera-0\" (UID: \"ddc18aa7-8599-40c5-a019-972c13162f83\") " pod="openstack/openstack-galera-0" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.694759 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hwh4v\" (UniqueName: \"kubernetes.io/projected/ddc18aa7-8599-40c5-a019-972c13162f83-kube-api-access-hwh4v\") pod \"openstack-galera-0\" (UID: \"ddc18aa7-8599-40c5-a019-972c13162f83\") " pod="openstack/openstack-galera-0" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.694846 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ddc18aa7-8599-40c5-a019-972c13162f83-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ddc18aa7-8599-40c5-a019-972c13162f83\") " pod="openstack/openstack-galera-0" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.694874 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddc18aa7-8599-40c5-a019-972c13162f83-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ddc18aa7-8599-40c5-a019-972c13162f83\") " pod="openstack/openstack-galera-0" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.694901 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ddc18aa7-8599-40c5-a019-972c13162f83-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ddc18aa7-8599-40c5-a019-972c13162f83\") " pod="openstack/openstack-galera-0" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.796239 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ddc18aa7-8599-40c5-a019-972c13162f83-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ddc18aa7-8599-40c5-a019-972c13162f83\") " pod="openstack/openstack-galera-0" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.796290 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddc18aa7-8599-40c5-a019-972c13162f83-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ddc18aa7-8599-40c5-a019-972c13162f83\") " pod="openstack/openstack-galera-0" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.796319 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ddc18aa7-8599-40c5-a019-972c13162f83-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ddc18aa7-8599-40c5-a019-972c13162f83\") " pod="openstack/openstack-galera-0" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.796382 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddc18aa7-8599-40c5-a019-972c13162f83-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ddc18aa7-8599-40c5-a019-972c13162f83\") " pod="openstack/openstack-galera-0" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.796415 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/ddc18aa7-8599-40c5-a019-972c13162f83-secrets\") pod \"openstack-galera-0\" (UID: \"ddc18aa7-8599-40c5-a019-972c13162f83\") " pod="openstack/openstack-galera-0" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.796440 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"ddc18aa7-8599-40c5-a019-972c13162f83\") " pod="openstack/openstack-galera-0" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.796480 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ddc18aa7-8599-40c5-a019-972c13162f83-kolla-config\") pod \"openstack-galera-0\" (UID: \"ddc18aa7-8599-40c5-a019-972c13162f83\") " pod="openstack/openstack-galera-0" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.796510 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ddc18aa7-8599-40c5-a019-972c13162f83-config-data-default\") pod \"openstack-galera-0\" (UID: \"ddc18aa7-8599-40c5-a019-972c13162f83\") " pod="openstack/openstack-galera-0" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.796538 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hwh4v\" (UniqueName: \"kubernetes.io/projected/ddc18aa7-8599-40c5-a019-972c13162f83-kube-api-access-hwh4v\") pod \"openstack-galera-0\" (UID: \"ddc18aa7-8599-40c5-a019-972c13162f83\") " pod="openstack/openstack-galera-0" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.797582 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/ddc18aa7-8599-40c5-a019-972c13162f83-kolla-config\") pod \"openstack-galera-0\" (UID: \"ddc18aa7-8599-40c5-a019-972c13162f83\") " pod="openstack/openstack-galera-0" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.797764 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/ddc18aa7-8599-40c5-a019-972c13162f83-config-data-default\") pod \"openstack-galera-0\" (UID: \"ddc18aa7-8599-40c5-a019-972c13162f83\") " pod="openstack/openstack-galera-0" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.798535 4967 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"ddc18aa7-8599-40c5-a019-972c13162f83\") device mount path \"/mnt/openstack/pv11\"" pod="openstack/openstack-galera-0" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.798945 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/ddc18aa7-8599-40c5-a019-972c13162f83-operator-scripts\") pod \"openstack-galera-0\" (UID: \"ddc18aa7-8599-40c5-a019-972c13162f83\") " pod="openstack/openstack-galera-0" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.799711 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/ddc18aa7-8599-40c5-a019-972c13162f83-secrets\") pod \"openstack-galera-0\" (UID: \"ddc18aa7-8599-40c5-a019-972c13162f83\") " pod="openstack/openstack-galera-0" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.800011 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ddc18aa7-8599-40c5-a019-972c13162f83-combined-ca-bundle\") pod \"openstack-galera-0\" (UID: \"ddc18aa7-8599-40c5-a019-972c13162f83\") " pod="openstack/openstack-galera-0" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.807834 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/ddc18aa7-8599-40c5-a019-972c13162f83-config-data-generated\") pod \"openstack-galera-0\" (UID: \"ddc18aa7-8599-40c5-a019-972c13162f83\") " pod="openstack/openstack-galera-0" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.815267 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hwh4v\" (UniqueName: \"kubernetes.io/projected/ddc18aa7-8599-40c5-a019-972c13162f83-kube-api-access-hwh4v\") pod \"openstack-galera-0\" (UID: \"ddc18aa7-8599-40c5-a019-972c13162f83\") " pod="openstack/openstack-galera-0" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.815632 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/ddc18aa7-8599-40c5-a019-972c13162f83-galera-tls-certs\") pod \"openstack-galera-0\" (UID: \"ddc18aa7-8599-40c5-a019-972c13162f83\") " pod="openstack/openstack-galera-0" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.828182 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage11-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage11-crc\") pod \"openstack-galera-0\" (UID: \"ddc18aa7-8599-40c5-a019-972c13162f83\") " pod="openstack/openstack-galera-0" Oct 11 04:06:02 crc kubenswrapper[4967]: I1011 04:06:02.914931 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-galera-0" Oct 11 04:06:03 crc kubenswrapper[4967]: I1011 04:06:03.683333 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-vc6g6" event={"ID":"b9b792fe-ab24-4587-a1f6-2ad194c9eb83","Type":"ContainerStarted","Data":"8e230d575816b5e3aa3bffe002ae98f8e2d23ac163f6e95713ce6d069fe9be0f"} Oct 11 04:06:03 crc kubenswrapper[4967]: I1011 04:06:03.974346 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 11 04:06:03 crc kubenswrapper[4967]: I1011 04:06:03.976259 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:03 crc kubenswrapper[4967]: I1011 04:06:03.979358 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-galera-openstack-cell1-svc" Oct 11 04:06:03 crc kubenswrapper[4967]: I1011 04:06:03.979490 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-config-data" Oct 11 04:06:03 crc kubenswrapper[4967]: I1011 04:06:03.979666 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-cell1-scripts" Oct 11 04:06:03 crc kubenswrapper[4967]: I1011 04:06:03.979861 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"galera-openstack-cell1-dockercfg-8m8jh" Oct 11 04:06:03 crc kubenswrapper[4967]: I1011 04:06:03.994261 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.117606 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f52a524-225c-4609-8cb0-0ec800e14aad-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"2f52a524-225c-4609-8cb0-0ec800e14aad\") " pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.117902 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/2f52a524-225c-4609-8cb0-0ec800e14aad-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"2f52a524-225c-4609-8cb0-0ec800e14aad\") " pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.117925 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"2f52a524-225c-4609-8cb0-0ec800e14aad\") " pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.117949 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2f52a524-225c-4609-8cb0-0ec800e14aad-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"2f52a524-225c-4609-8cb0-0ec800e14aad\") " pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.118016 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f52a524-225c-4609-8cb0-0ec800e14aad-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"2f52a524-225c-4609-8cb0-0ec800e14aad\") " pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.118036 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f52a524-225c-4609-8cb0-0ec800e14aad-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"2f52a524-225c-4609-8cb0-0ec800e14aad\") " pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.118062 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-snkg8\" (UniqueName: \"kubernetes.io/projected/2f52a524-225c-4609-8cb0-0ec800e14aad-kube-api-access-snkg8\") pod \"openstack-cell1-galera-0\" (UID: \"2f52a524-225c-4609-8cb0-0ec800e14aad\") " pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.118111 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2f52a524-225c-4609-8cb0-0ec800e14aad-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"2f52a524-225c-4609-8cb0-0ec800e14aad\") " pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.118155 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2f52a524-225c-4609-8cb0-0ec800e14aad-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"2f52a524-225c-4609-8cb0-0ec800e14aad\") " pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.201811 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/memcached-0"] Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.203446 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.208855 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.244975 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-memcached-svc" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.245023 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"memcached-config-data" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.245157 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"memcached-memcached-dockercfg-5lfnd" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.245358 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f52a524-225c-4609-8cb0-0ec800e14aad-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"2f52a524-225c-4609-8cb0-0ec800e14aad\") " pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.245402 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/2f52a524-225c-4609-8cb0-0ec800e14aad-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"2f52a524-225c-4609-8cb0-0ec800e14aad\") " pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.245422 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"2f52a524-225c-4609-8cb0-0ec800e14aad\") " pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.245445 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2f52a524-225c-4609-8cb0-0ec800e14aad-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"2f52a524-225c-4609-8cb0-0ec800e14aad\") " pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.245475 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f52a524-225c-4609-8cb0-0ec800e14aad-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"2f52a524-225c-4609-8cb0-0ec800e14aad\") " pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.245491 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f52a524-225c-4609-8cb0-0ec800e14aad-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"2f52a524-225c-4609-8cb0-0ec800e14aad\") " pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.245507 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-snkg8\" (UniqueName: \"kubernetes.io/projected/2f52a524-225c-4609-8cb0-0ec800e14aad-kube-api-access-snkg8\") pod \"openstack-cell1-galera-0\" (UID: \"2f52a524-225c-4609-8cb0-0ec800e14aad\") " pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.245528 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2f52a524-225c-4609-8cb0-0ec800e14aad-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"2f52a524-225c-4609-8cb0-0ec800e14aad\") " pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.245575 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2f52a524-225c-4609-8cb0-0ec800e14aad-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"2f52a524-225c-4609-8cb0-0ec800e14aad\") " pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.245954 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-generated\" (UniqueName: \"kubernetes.io/empty-dir/2f52a524-225c-4609-8cb0-0ec800e14aad-config-data-generated\") pod \"openstack-cell1-galera-0\" (UID: \"2f52a524-225c-4609-8cb0-0ec800e14aad\") " pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.246355 4967 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"2f52a524-225c-4609-8cb0-0ec800e14aad\") device mount path \"/mnt/openstack/pv08\"" pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.246399 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/2f52a524-225c-4609-8cb0-0ec800e14aad-kolla-config\") pod \"openstack-cell1-galera-0\" (UID: \"2f52a524-225c-4609-8cb0-0ec800e14aad\") " pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.246753 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-default\" (UniqueName: \"kubernetes.io/configmap/2f52a524-225c-4609-8cb0-0ec800e14aad-config-data-default\") pod \"openstack-cell1-galera-0\" (UID: \"2f52a524-225c-4609-8cb0-0ec800e14aad\") " pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.247584 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"operator-scripts\" (UniqueName: \"kubernetes.io/configmap/2f52a524-225c-4609-8cb0-0ec800e14aad-operator-scripts\") pod \"openstack-cell1-galera-0\" (UID: \"2f52a524-225c-4609-8cb0-0ec800e14aad\") " pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.253776 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2f52a524-225c-4609-8cb0-0ec800e14aad-combined-ca-bundle\") pod \"openstack-cell1-galera-0\" (UID: \"2f52a524-225c-4609-8cb0-0ec800e14aad\") " pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.253901 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"galera-tls-certs\" (UniqueName: \"kubernetes.io/secret/2f52a524-225c-4609-8cb0-0ec800e14aad-galera-tls-certs\") pod \"openstack-cell1-galera-0\" (UID: \"2f52a524-225c-4609-8cb0-0ec800e14aad\") " pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.262755 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secrets\" (UniqueName: \"kubernetes.io/secret/2f52a524-225c-4609-8cb0-0ec800e14aad-secrets\") pod \"openstack-cell1-galera-0\" (UID: \"2f52a524-225c-4609-8cb0-0ec800e14aad\") " pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.269585 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-snkg8\" (UniqueName: \"kubernetes.io/projected/2f52a524-225c-4609-8cb0-0ec800e14aad-kube-api-access-snkg8\") pod \"openstack-cell1-galera-0\" (UID: \"2f52a524-225c-4609-8cb0-0ec800e14aad\") " pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.276212 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage08-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage08-crc\") pod \"openstack-cell1-galera-0\" (UID: \"2f52a524-225c-4609-8cb0-0ec800e14aad\") " pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.303041 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.347060 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vsjqd\" (UniqueName: \"kubernetes.io/projected/10ea033f-281c-4519-8da7-93e598f08b11-kube-api-access-vsjqd\") pod \"memcached-0\" (UID: \"10ea033f-281c-4519-8da7-93e598f08b11\") " pod="openstack/memcached-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.347117 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/10ea033f-281c-4519-8da7-93e598f08b11-kolla-config\") pod \"memcached-0\" (UID: \"10ea033f-281c-4519-8da7-93e598f08b11\") " pod="openstack/memcached-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.347141 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/10ea033f-281c-4519-8da7-93e598f08b11-memcached-tls-certs\") pod \"memcached-0\" (UID: \"10ea033f-281c-4519-8da7-93e598f08b11\") " pod="openstack/memcached-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.347158 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10ea033f-281c-4519-8da7-93e598f08b11-combined-ca-bundle\") pod \"memcached-0\" (UID: \"10ea033f-281c-4519-8da7-93e598f08b11\") " pod="openstack/memcached-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.347232 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/10ea033f-281c-4519-8da7-93e598f08b11-config-data\") pod \"memcached-0\" (UID: \"10ea033f-281c-4519-8da7-93e598f08b11\") " pod="openstack/memcached-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.448585 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/10ea033f-281c-4519-8da7-93e598f08b11-config-data\") pod \"memcached-0\" (UID: \"10ea033f-281c-4519-8da7-93e598f08b11\") " pod="openstack/memcached-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.448648 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vsjqd\" (UniqueName: \"kubernetes.io/projected/10ea033f-281c-4519-8da7-93e598f08b11-kube-api-access-vsjqd\") pod \"memcached-0\" (UID: \"10ea033f-281c-4519-8da7-93e598f08b11\") " pod="openstack/memcached-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.448692 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/10ea033f-281c-4519-8da7-93e598f08b11-kolla-config\") pod \"memcached-0\" (UID: \"10ea033f-281c-4519-8da7-93e598f08b11\") " pod="openstack/memcached-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.448728 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/10ea033f-281c-4519-8da7-93e598f08b11-memcached-tls-certs\") pod \"memcached-0\" (UID: \"10ea033f-281c-4519-8da7-93e598f08b11\") " pod="openstack/memcached-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.448754 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10ea033f-281c-4519-8da7-93e598f08b11-combined-ca-bundle\") pod \"memcached-0\" (UID: \"10ea033f-281c-4519-8da7-93e598f08b11\") " pod="openstack/memcached-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.449608 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kolla-config\" (UniqueName: \"kubernetes.io/configmap/10ea033f-281c-4519-8da7-93e598f08b11-kolla-config\") pod \"memcached-0\" (UID: \"10ea033f-281c-4519-8da7-93e598f08b11\") " pod="openstack/memcached-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.450125 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/10ea033f-281c-4519-8da7-93e598f08b11-config-data\") pod \"memcached-0\" (UID: \"10ea033f-281c-4519-8da7-93e598f08b11\") " pod="openstack/memcached-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.453449 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"memcached-tls-certs\" (UniqueName: \"kubernetes.io/secret/10ea033f-281c-4519-8da7-93e598f08b11-memcached-tls-certs\") pod \"memcached-0\" (UID: \"10ea033f-281c-4519-8da7-93e598f08b11\") " pod="openstack/memcached-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.459270 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/10ea033f-281c-4519-8da7-93e598f08b11-combined-ca-bundle\") pod \"memcached-0\" (UID: \"10ea033f-281c-4519-8da7-93e598f08b11\") " pod="openstack/memcached-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.463878 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vsjqd\" (UniqueName: \"kubernetes.io/projected/10ea033f-281c-4519-8da7-93e598f08b11-kube-api-access-vsjqd\") pod \"memcached-0\" (UID: \"10ea033f-281c-4519-8da7-93e598f08b11\") " pod="openstack/memcached-0" Oct 11 04:06:04 crc kubenswrapper[4967]: I1011 04:06:04.617462 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/memcached-0" Oct 11 04:06:06 crc kubenswrapper[4967]: I1011 04:06:06.095623 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 11 04:06:06 crc kubenswrapper[4967]: I1011 04:06:06.097044 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 11 04:06:06 crc kubenswrapper[4967]: I1011 04:06:06.100644 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"telemetry-ceilometer-dockercfg-gg9fl" Oct 11 04:06:06 crc kubenswrapper[4967]: I1011 04:06:06.106176 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 11 04:06:06 crc kubenswrapper[4967]: I1011 04:06:06.276694 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4t756\" (UniqueName: \"kubernetes.io/projected/3d120d8c-893d-49d9-8649-453997725ff2-kube-api-access-4t756\") pod \"kube-state-metrics-0\" (UID: \"3d120d8c-893d-49d9-8649-453997725ff2\") " pod="openstack/kube-state-metrics-0" Oct 11 04:06:06 crc kubenswrapper[4967]: I1011 04:06:06.377740 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4t756\" (UniqueName: \"kubernetes.io/projected/3d120d8c-893d-49d9-8649-453997725ff2-kube-api-access-4t756\") pod \"kube-state-metrics-0\" (UID: \"3d120d8c-893d-49d9-8649-453997725ff2\") " pod="openstack/kube-state-metrics-0" Oct 11 04:06:06 crc kubenswrapper[4967]: I1011 04:06:06.400009 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4t756\" (UniqueName: \"kubernetes.io/projected/3d120d8c-893d-49d9-8649-453997725ff2-kube-api-access-4t756\") pod \"kube-state-metrics-0\" (UID: \"3d120d8c-893d-49d9-8649-453997725ff2\") " pod="openstack/kube-state-metrics-0" Oct 11 04:06:06 crc kubenswrapper[4967]: I1011 04:06:06.415810 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.537348 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.540492 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.544557 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-scripts" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.544899 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-sb-config" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.545134 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovn-metrics" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.545229 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-sb-ovndbs" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.545234 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-sb-dockercfg-7bctd" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.572780 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.671468 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca\") " pod="openstack/ovsdbserver-sb-0" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.671954 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca\") " pod="openstack/ovsdbserver-sb-0" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.671993 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca\") " pod="openstack/ovsdbserver-sb-0" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.672162 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca\") " pod="openstack/ovsdbserver-sb-0" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.672221 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca\") " pod="openstack/ovsdbserver-sb-0" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.672277 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kczxk\" (UniqueName: \"kubernetes.io/projected/36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca-kube-api-access-kczxk\") pod \"ovsdbserver-sb-0\" (UID: \"36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca\") " pod="openstack/ovsdbserver-sb-0" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.672329 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca\") " pod="openstack/ovsdbserver-sb-0" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.672409 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca-config\") pod \"ovsdbserver-sb-0\" (UID: \"36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca\") " pod="openstack/ovsdbserver-sb-0" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.774862 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca\") " pod="openstack/ovsdbserver-sb-0" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.774976 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca\") " pod="openstack/ovsdbserver-sb-0" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.775016 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca\") " pod="openstack/ovsdbserver-sb-0" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.775118 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca\") " pod="openstack/ovsdbserver-sb-0" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.775185 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca\") " pod="openstack/ovsdbserver-sb-0" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.775221 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kczxk\" (UniqueName: \"kubernetes.io/projected/36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca-kube-api-access-kczxk\") pod \"ovsdbserver-sb-0\" (UID: \"36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca\") " pod="openstack/ovsdbserver-sb-0" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.775298 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca\") " pod="openstack/ovsdbserver-sb-0" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.775389 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca-config\") pod \"ovsdbserver-sb-0\" (UID: \"36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca\") " pod="openstack/ovsdbserver-sb-0" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.775876 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca-ovsdb-rundir\") pod \"ovsdbserver-sb-0\" (UID: \"36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca\") " pod="openstack/ovsdbserver-sb-0" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.777898 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca-scripts\") pod \"ovsdbserver-sb-0\" (UID: \"36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca\") " pod="openstack/ovsdbserver-sb-0" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.779162 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca-config\") pod \"ovsdbserver-sb-0\" (UID: \"36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca\") " pod="openstack/ovsdbserver-sb-0" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.786340 4967 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca\") device mount path \"/mnt/openstack/pv12\"" pod="openstack/ovsdbserver-sb-0" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.792724 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca-metrics-certs-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca\") " pod="openstack/ovsdbserver-sb-0" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.794862 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb-tls-certs\" (UniqueName: \"kubernetes.io/secret/36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca-ovsdbserver-sb-tls-certs\") pod \"ovsdbserver-sb-0\" (UID: \"36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca\") " pod="openstack/ovsdbserver-sb-0" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.801215 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-wvrnq"] Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.802620 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-wvrnq" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.805337 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-ovs-7b5rp"] Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.807142 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-7b5rp" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.816435 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-scripts" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.817050 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncontroller-ovncontroller-dockercfg-vrdj7" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.817183 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovncontroller-ovndbs" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.816797 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca-combined-ca-bundle\") pod \"ovsdbserver-sb-0\" (UID: \"36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca\") " pod="openstack/ovsdbserver-sb-0" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.827953 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kczxk\" (UniqueName: \"kubernetes.io/projected/36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca-kube-api-access-kczxk\") pod \"ovsdbserver-sb-0\" (UID: \"36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca\") " pod="openstack/ovsdbserver-sb-0" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.844617 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-wvrnq"] Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.872436 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-7b5rp"] Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.880944 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage12-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage12-crc\") pod \"ovsdbserver-sb-0\" (UID: \"36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca\") " pod="openstack/ovsdbserver-sb-0" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.904551 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5vbgv"] Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.978705 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/acad8def-db3f-40a3-a95d-28b7503d4aae-ovn-controller-tls-certs\") pod \"ovn-controller-wvrnq\" (UID: \"acad8def-db3f-40a3-a95d-28b7503d4aae\") " pod="openstack/ovn-controller-wvrnq" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.978795 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/acad8def-db3f-40a3-a95d-28b7503d4aae-var-run\") pod \"ovn-controller-wvrnq\" (UID: \"acad8def-db3f-40a3-a95d-28b7503d4aae\") " pod="openstack/ovn-controller-wvrnq" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.978830 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acad8def-db3f-40a3-a95d-28b7503d4aae-combined-ca-bundle\") pod \"ovn-controller-wvrnq\" (UID: \"acad8def-db3f-40a3-a95d-28b7503d4aae\") " pod="openstack/ovn-controller-wvrnq" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.978863 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-btj7m\" (UniqueName: \"kubernetes.io/projected/acad8def-db3f-40a3-a95d-28b7503d4aae-kube-api-access-btj7m\") pod \"ovn-controller-wvrnq\" (UID: \"acad8def-db3f-40a3-a95d-28b7503d4aae\") " pod="openstack/ovn-controller-wvrnq" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.978898 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/acad8def-db3f-40a3-a95d-28b7503d4aae-scripts\") pod \"ovn-controller-wvrnq\" (UID: \"acad8def-db3f-40a3-a95d-28b7503d4aae\") " pod="openstack/ovn-controller-wvrnq" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.978941 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20-etc-ovs\") pod \"ovn-controller-ovs-7b5rp\" (UID: \"4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20\") " pod="openstack/ovn-controller-ovs-7b5rp" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.978969 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20-var-log\") pod \"ovn-controller-ovs-7b5rp\" (UID: \"4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20\") " pod="openstack/ovn-controller-ovs-7b5rp" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.978989 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20-var-run\") pod \"ovn-controller-ovs-7b5rp\" (UID: \"4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20\") " pod="openstack/ovn-controller-ovs-7b5rp" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.979014 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20-scripts\") pod \"ovn-controller-ovs-7b5rp\" (UID: \"4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20\") " pod="openstack/ovn-controller-ovs-7b5rp" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.979043 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/acad8def-db3f-40a3-a95d-28b7503d4aae-var-run-ovn\") pod \"ovn-controller-wvrnq\" (UID: \"acad8def-db3f-40a3-a95d-28b7503d4aae\") " pod="openstack/ovn-controller-wvrnq" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.979084 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20-var-lib\") pod \"ovn-controller-ovs-7b5rp\" (UID: \"4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20\") " pod="openstack/ovn-controller-ovs-7b5rp" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.979116 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cp9sr\" (UniqueName: \"kubernetes.io/projected/4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20-kube-api-access-cp9sr\") pod \"ovn-controller-ovs-7b5rp\" (UID: \"4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20\") " pod="openstack/ovn-controller-ovs-7b5rp" Oct 11 04:06:10 crc kubenswrapper[4967]: I1011 04:06:10.979169 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/acad8def-db3f-40a3-a95d-28b7503d4aae-var-log-ovn\") pod \"ovn-controller-wvrnq\" (UID: \"acad8def-db3f-40a3-a95d-28b7503d4aae\") " pod="openstack/ovn-controller-wvrnq" Oct 11 04:06:11 crc kubenswrapper[4967]: I1011 04:06:11.080789 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/acad8def-db3f-40a3-a95d-28b7503d4aae-var-log-ovn\") pod \"ovn-controller-wvrnq\" (UID: \"acad8def-db3f-40a3-a95d-28b7503d4aae\") " pod="openstack/ovn-controller-wvrnq" Oct 11 04:06:11 crc kubenswrapper[4967]: I1011 04:06:11.080856 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/acad8def-db3f-40a3-a95d-28b7503d4aae-ovn-controller-tls-certs\") pod \"ovn-controller-wvrnq\" (UID: \"acad8def-db3f-40a3-a95d-28b7503d4aae\") " pod="openstack/ovn-controller-wvrnq" Oct 11 04:06:11 crc kubenswrapper[4967]: I1011 04:06:11.080877 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/acad8def-db3f-40a3-a95d-28b7503d4aae-var-run\") pod \"ovn-controller-wvrnq\" (UID: \"acad8def-db3f-40a3-a95d-28b7503d4aae\") " pod="openstack/ovn-controller-wvrnq" Oct 11 04:06:11 crc kubenswrapper[4967]: I1011 04:06:11.080898 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acad8def-db3f-40a3-a95d-28b7503d4aae-combined-ca-bundle\") pod \"ovn-controller-wvrnq\" (UID: \"acad8def-db3f-40a3-a95d-28b7503d4aae\") " pod="openstack/ovn-controller-wvrnq" Oct 11 04:06:11 crc kubenswrapper[4967]: I1011 04:06:11.080924 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-btj7m\" (UniqueName: \"kubernetes.io/projected/acad8def-db3f-40a3-a95d-28b7503d4aae-kube-api-access-btj7m\") pod \"ovn-controller-wvrnq\" (UID: \"acad8def-db3f-40a3-a95d-28b7503d4aae\") " pod="openstack/ovn-controller-wvrnq" Oct 11 04:06:11 crc kubenswrapper[4967]: I1011 04:06:11.080952 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/acad8def-db3f-40a3-a95d-28b7503d4aae-scripts\") pod \"ovn-controller-wvrnq\" (UID: \"acad8def-db3f-40a3-a95d-28b7503d4aae\") " pod="openstack/ovn-controller-wvrnq" Oct 11 04:06:11 crc kubenswrapper[4967]: I1011 04:06:11.080999 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20-etc-ovs\") pod \"ovn-controller-ovs-7b5rp\" (UID: \"4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20\") " pod="openstack/ovn-controller-ovs-7b5rp" Oct 11 04:06:11 crc kubenswrapper[4967]: I1011 04:06:11.081022 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20-var-log\") pod \"ovn-controller-ovs-7b5rp\" (UID: \"4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20\") " pod="openstack/ovn-controller-ovs-7b5rp" Oct 11 04:06:11 crc kubenswrapper[4967]: I1011 04:06:11.081040 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20-var-run\") pod \"ovn-controller-ovs-7b5rp\" (UID: \"4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20\") " pod="openstack/ovn-controller-ovs-7b5rp" Oct 11 04:06:11 crc kubenswrapper[4967]: I1011 04:06:11.081057 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20-scripts\") pod \"ovn-controller-ovs-7b5rp\" (UID: \"4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20\") " pod="openstack/ovn-controller-ovs-7b5rp" Oct 11 04:06:11 crc kubenswrapper[4967]: I1011 04:06:11.081094 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/acad8def-db3f-40a3-a95d-28b7503d4aae-var-run-ovn\") pod \"ovn-controller-wvrnq\" (UID: \"acad8def-db3f-40a3-a95d-28b7503d4aae\") " pod="openstack/ovn-controller-wvrnq" Oct 11 04:06:11 crc kubenswrapper[4967]: I1011 04:06:11.081114 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20-var-lib\") pod \"ovn-controller-ovs-7b5rp\" (UID: \"4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20\") " pod="openstack/ovn-controller-ovs-7b5rp" Oct 11 04:06:11 crc kubenswrapper[4967]: I1011 04:06:11.081136 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cp9sr\" (UniqueName: \"kubernetes.io/projected/4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20-kube-api-access-cp9sr\") pod \"ovn-controller-ovs-7b5rp\" (UID: \"4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20\") " pod="openstack/ovn-controller-ovs-7b5rp" Oct 11 04:06:11 crc kubenswrapper[4967]: I1011 04:06:11.081986 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/acad8def-db3f-40a3-a95d-28b7503d4aae-var-log-ovn\") pod \"ovn-controller-wvrnq\" (UID: \"acad8def-db3f-40a3-a95d-28b7503d4aae\") " pod="openstack/ovn-controller-wvrnq" Oct 11 04:06:11 crc kubenswrapper[4967]: I1011 04:06:11.082103 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ovs\" (UniqueName: \"kubernetes.io/host-path/4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20-etc-ovs\") pod \"ovn-controller-ovs-7b5rp\" (UID: \"4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20\") " pod="openstack/ovn-controller-ovs-7b5rp" Oct 11 04:06:11 crc kubenswrapper[4967]: I1011 04:06:11.082119 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20-var-log\") pod \"ovn-controller-ovs-7b5rp\" (UID: \"4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20\") " pod="openstack/ovn-controller-ovs-7b5rp" Oct 11 04:06:11 crc kubenswrapper[4967]: I1011 04:06:11.082213 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20-var-run\") pod \"ovn-controller-ovs-7b5rp\" (UID: \"4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20\") " pod="openstack/ovn-controller-ovs-7b5rp" Oct 11 04:06:11 crc kubenswrapper[4967]: I1011 04:06:11.083171 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/acad8def-db3f-40a3-a95d-28b7503d4aae-var-run\") pod \"ovn-controller-wvrnq\" (UID: \"acad8def-db3f-40a3-a95d-28b7503d4aae\") " pod="openstack/ovn-controller-wvrnq" Oct 11 04:06:11 crc kubenswrapper[4967]: I1011 04:06:11.083398 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/acad8def-db3f-40a3-a95d-28b7503d4aae-var-run-ovn\") pod \"ovn-controller-wvrnq\" (UID: \"acad8def-db3f-40a3-a95d-28b7503d4aae\") " pod="openstack/ovn-controller-wvrnq" Oct 11 04:06:11 crc kubenswrapper[4967]: I1011 04:06:11.083496 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib\" (UniqueName: \"kubernetes.io/host-path/4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20-var-lib\") pod \"ovn-controller-ovs-7b5rp\" (UID: \"4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20\") " pod="openstack/ovn-controller-ovs-7b5rp" Oct 11 04:06:11 crc kubenswrapper[4967]: I1011 04:06:11.085577 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/acad8def-db3f-40a3-a95d-28b7503d4aae-scripts\") pod \"ovn-controller-wvrnq\" (UID: \"acad8def-db3f-40a3-a95d-28b7503d4aae\") " pod="openstack/ovn-controller-wvrnq" Oct 11 04:06:11 crc kubenswrapper[4967]: I1011 04:06:11.085621 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/acad8def-db3f-40a3-a95d-28b7503d4aae-combined-ca-bundle\") pod \"ovn-controller-wvrnq\" (UID: \"acad8def-db3f-40a3-a95d-28b7503d4aae\") " pod="openstack/ovn-controller-wvrnq" Oct 11 04:06:11 crc kubenswrapper[4967]: I1011 04:06:11.086768 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20-scripts\") pod \"ovn-controller-ovs-7b5rp\" (UID: \"4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20\") " pod="openstack/ovn-controller-ovs-7b5rp" Oct 11 04:06:11 crc kubenswrapper[4967]: I1011 04:06:11.087246 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-controller-tls-certs\" (UniqueName: \"kubernetes.io/secret/acad8def-db3f-40a3-a95d-28b7503d4aae-ovn-controller-tls-certs\") pod \"ovn-controller-wvrnq\" (UID: \"acad8def-db3f-40a3-a95d-28b7503d4aae\") " pod="openstack/ovn-controller-wvrnq" Oct 11 04:06:11 crc kubenswrapper[4967]: I1011 04:06:11.100928 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-btj7m\" (UniqueName: \"kubernetes.io/projected/acad8def-db3f-40a3-a95d-28b7503d4aae-kube-api-access-btj7m\") pod \"ovn-controller-wvrnq\" (UID: \"acad8def-db3f-40a3-a95d-28b7503d4aae\") " pod="openstack/ovn-controller-wvrnq" Oct 11 04:06:11 crc kubenswrapper[4967]: I1011 04:06:11.101440 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cp9sr\" (UniqueName: \"kubernetes.io/projected/4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20-kube-api-access-cp9sr\") pod \"ovn-controller-ovs-7b5rp\" (UID: \"4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20\") " pod="openstack/ovn-controller-ovs-7b5rp" Oct 11 04:06:11 crc kubenswrapper[4967]: I1011 04:06:11.165388 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-sb-0" Oct 11 04:06:11 crc kubenswrapper[4967]: I1011 04:06:11.198157 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-wvrnq" Oct 11 04:06:11 crc kubenswrapper[4967]: I1011 04:06:11.199520 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-ovs-7b5rp" Oct 11 04:06:11 crc kubenswrapper[4967]: W1011 04:06:11.520474 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod803e1df8_1759_49e3_83c8_abb1abf6fff4.slice/crio-b2f782f032c59630f6f26294695df03be2b32b713cb7a24fc687666e9fc9062c WatchSource:0}: Error finding container b2f782f032c59630f6f26294695df03be2b32b713cb7a24fc687666e9fc9062c: Status 404 returned error can't find the container with id b2f782f032c59630f6f26294695df03be2b32b713cb7a24fc687666e9fc9062c Oct 11 04:06:11 crc kubenswrapper[4967]: I1011 04:06:11.525856 4967 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 11 04:06:11 crc kubenswrapper[4967]: E1011 04:06:11.531845 4967 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 11 04:06:11 crc kubenswrapper[4967]: E1011 04:06:11.532034 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:nffh5bdhf4h5f8h79h55h77h58fh56dh7bh6fh578hbch55dh68h56bhd9h65dh57ch658hc9h566h666h688h58h65dh684h5d7h6ch575h5d6h88q,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-gqpp9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-675f4bcbfc-lxnrr_openstack(f980d5dd-67f3-4714-a626-d521efde4916): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 11 04:06:11 crc kubenswrapper[4967]: E1011 04:06:11.533538 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-675f4bcbfc-lxnrr" podUID="f980d5dd-67f3-4714-a626-d521efde4916" Oct 11 04:06:11 crc kubenswrapper[4967]: E1011 04:06:11.573447 4967 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified" Oct 11 04:06:11 crc kubenswrapper[4967]: E1011 04:06:11.573915 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:init,Image:quay.io/podified-antelope-centos9/openstack-neutron-server:current-podified,Command:[/bin/bash],Args:[-c dnsmasq --interface=* --conf-dir=/etc/dnsmasq.d --hostsdir=/etc/dnsmasq.d/hosts --keep-in-foreground --log-debug --bind-interfaces --listen-address=$(POD_IP) --port 5353 --log-facility=- --no-hosts --domain-needed --no-resolv --bogus-priv --log-queries --test],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:CONFIG_HASH,Value:ndfhb5h667h568h584h5f9h58dh565h664h587h597h577h64bh5c4h66fh647hbdh68ch5c5h68dh686h5f7h64hd7hc6h55fh57bh98h57fh87h5fh57fq,ValueFrom:nil,},EnvVar{Name:POD_IP,Value:,ValueFrom:&EnvVarSource{FieldRef:&ObjectFieldSelector{APIVersion:v1,FieldPath:status.podIP,},ResourceFieldRef:nil,ConfigMapKeyRef:nil,SecretKeyRef:nil,},},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:config,ReadOnly:true,MountPath:/etc/dnsmasq.d/config.cfg,SubPath:dns,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:dns-svc,ReadOnly:true,MountPath:/etc/dnsmasq.d/hosts/dns-svc,SubPath:dns-svc,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-4s4ft,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000650000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:&SeccompProfile{Type:RuntimeDefault,LocalhostProfile:nil,},AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod dnsmasq-dns-78dd6ddcc-8wsqp_openstack(854c2370-c43a-4c50-99ab-5d935456c2ae): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 11 04:06:11 crc kubenswrapper[4967]: E1011 04:06:11.575646 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"init\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/dnsmasq-dns-78dd6ddcc-8wsqp" podUID="854c2370-c43a-4c50-99ab-5d935456c2ae" Oct 11 04:06:11 crc kubenswrapper[4967]: I1011 04:06:11.762255 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-5vbgv" event={"ID":"803e1df8-1759-49e3-83c8-abb1abf6fff4","Type":"ContainerStarted","Data":"b2f782f032c59630f6f26294695df03be2b32b713cb7a24fc687666e9fc9062c"} Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.347280 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-8wsqp" Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.373146 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/memcached-0"] Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.468551 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-lxnrr" Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.508571 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/854c2370-c43a-4c50-99ab-5d935456c2ae-config\") pod \"854c2370-c43a-4c50-99ab-5d935456c2ae\" (UID: \"854c2370-c43a-4c50-99ab-5d935456c2ae\") " Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.508614 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4s4ft\" (UniqueName: \"kubernetes.io/projected/854c2370-c43a-4c50-99ab-5d935456c2ae-kube-api-access-4s4ft\") pod \"854c2370-c43a-4c50-99ab-5d935456c2ae\" (UID: \"854c2370-c43a-4c50-99ab-5d935456c2ae\") " Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.508677 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/854c2370-c43a-4c50-99ab-5d935456c2ae-dns-svc\") pod \"854c2370-c43a-4c50-99ab-5d935456c2ae\" (UID: \"854c2370-c43a-4c50-99ab-5d935456c2ae\") " Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.509216 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/854c2370-c43a-4c50-99ab-5d935456c2ae-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "854c2370-c43a-4c50-99ab-5d935456c2ae" (UID: "854c2370-c43a-4c50-99ab-5d935456c2ae"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.509712 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/854c2370-c43a-4c50-99ab-5d935456c2ae-config" (OuterVolumeSpecName: "config") pod "854c2370-c43a-4c50-99ab-5d935456c2ae" (UID: "854c2370-c43a-4c50-99ab-5d935456c2ae"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.510188 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/854c2370-c43a-4c50-99ab-5d935456c2ae-config\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.510210 4967 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/854c2370-c43a-4c50-99ab-5d935456c2ae-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.513053 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/854c2370-c43a-4c50-99ab-5d935456c2ae-kube-api-access-4s4ft" (OuterVolumeSpecName: "kube-api-access-4s4ft") pod "854c2370-c43a-4c50-99ab-5d935456c2ae" (UID: "854c2370-c43a-4c50-99ab-5d935456c2ae"). InnerVolumeSpecName "kube-api-access-4s4ft". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.611034 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f980d5dd-67f3-4714-a626-d521efde4916-config\") pod \"f980d5dd-67f3-4714-a626-d521efde4916\" (UID: \"f980d5dd-67f3-4714-a626-d521efde4916\") " Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.611172 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqpp9\" (UniqueName: \"kubernetes.io/projected/f980d5dd-67f3-4714-a626-d521efde4916-kube-api-access-gqpp9\") pod \"f980d5dd-67f3-4714-a626-d521efde4916\" (UID: \"f980d5dd-67f3-4714-a626-d521efde4916\") " Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.611441 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4s4ft\" (UniqueName: \"kubernetes.io/projected/854c2370-c43a-4c50-99ab-5d935456c2ae-kube-api-access-4s4ft\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.611447 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f980d5dd-67f3-4714-a626-d521efde4916-config" (OuterVolumeSpecName: "config") pod "f980d5dd-67f3-4714-a626-d521efde4916" (UID: "f980d5dd-67f3-4714-a626-d521efde4916"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.611843 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-cell1-galera-0"] Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.614212 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f980d5dd-67f3-4714-a626-d521efde4916-kube-api-access-gqpp9" (OuterVolumeSpecName: "kube-api-access-gqpp9") pod "f980d5dd-67f3-4714-a626-d521efde4916" (UID: "f980d5dd-67f3-4714-a626-d521efde4916"). InnerVolumeSpecName "kube-api-access-gqpp9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.623968 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 11 04:06:12 crc kubenswrapper[4967]: W1011 04:06:12.630333 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2f52a524_225c_4609_8cb0_0ec800e14aad.slice/crio-08c6fac38a08a5fdb58e625149872355ce37f196678173626a9383f40610965f WatchSource:0}: Error finding container 08c6fac38a08a5fdb58e625149872355ce37f196678173626a9383f40610965f: Status 404 returned error can't find the container with id 08c6fac38a08a5fdb58e625149872355ce37f196678173626a9383f40610965f Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.639899 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstack-galera-0"] Oct 11 04:06:12 crc kubenswrapper[4967]: W1011 04:06:12.651793 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podacad8def_db3f_40a3_a95d_28b7503d4aae.slice/crio-7b727eae69cba0775bedffce79e00fa5dd094bb6d40a4ab13c86abbfcfc4ab9e WatchSource:0}: Error finding container 7b727eae69cba0775bedffce79e00fa5dd094bb6d40a4ab13c86abbfcfc4ab9e: Status 404 returned error can't find the container with id 7b727eae69cba0775bedffce79e00fa5dd094bb6d40a4ab13c86abbfcfc4ab9e Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.651954 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-wvrnq"] Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.676166 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 11 04:06:12 crc kubenswrapper[4967]: W1011 04:06:12.684137 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d923c19_a45c_4cb7_9f19_e770c5fe66ed.slice/crio-b5f4e5a5434cd9517206736b24fd70f5ca589f42fce0f05d46488df83b0a74e0 WatchSource:0}: Error finding container b5f4e5a5434cd9517206736b24fd70f5ca589f42fce0f05d46488df83b0a74e0: Status 404 returned error can't find the container with id b5f4e5a5434cd9517206736b24fd70f5ca589f42fce0f05d46488df83b0a74e0 Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.697785 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.714868 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/f980d5dd-67f3-4714-a626-d521efde4916-config\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.714908 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqpp9\" (UniqueName: \"kubernetes.io/projected/f980d5dd-67f3-4714-a626-d521efde4916-kube-api-access-gqpp9\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.760565 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-sb-0"] Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.782425 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"10ea033f-281c-4519-8da7-93e598f08b11","Type":"ContainerStarted","Data":"aac52b6bbc79c0ce61600a38922eb45e03de99581c3a95a70b82eab895d854cb"} Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.786526 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ddc18aa7-8599-40c5-a019-972c13162f83","Type":"ContainerStarted","Data":"71e8e84041f564694db7bcec8b7a34db60ef377c6d6579b9ce301252ee6f7702"} Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.787490 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-wvrnq" event={"ID":"acad8def-db3f-40a3-a95d-28b7503d4aae","Type":"ContainerStarted","Data":"7b727eae69cba0775bedffce79e00fa5dd094bb6d40a4ab13c86abbfcfc4ab9e"} Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.788872 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0c315653-59d1-429f-be9c-c1ac826cff49","Type":"ContainerStarted","Data":"cfe9b07ca4a82285c51651ed38d8d66a562ad2394d3de32e4db0214795030210"} Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.790202 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-675f4bcbfc-lxnrr" Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.791216 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-675f4bcbfc-lxnrr" event={"ID":"f980d5dd-67f3-4714-a626-d521efde4916","Type":"ContainerDied","Data":"846dde0fc4421a635bdd8a534ba89c85cf76194c7b4475a1fe486b0aa603fb23"} Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.796215 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3d923c19-a45c-4cb7-9f19-e770c5fe66ed","Type":"ContainerStarted","Data":"b5f4e5a5434cd9517206736b24fd70f5ca589f42fce0f05d46488df83b0a74e0"} Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.798286 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3d120d8c-893d-49d9-8649-453997725ff2","Type":"ContainerStarted","Data":"18078965abda972022268296e54fa6994f5e9c47809a5360b54b182a2cc6ceb1"} Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.802431 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-78dd6ddcc-8wsqp" event={"ID":"854c2370-c43a-4c50-99ab-5d935456c2ae","Type":"ContainerDied","Data":"a6bc70e9455375f1ba08ce5d78aacef5cee9a5ce589d9945d3399b44feeb1472"} Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.802508 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-78dd6ddcc-8wsqp" Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.809162 4967 generic.go:334] "Generic (PLEG): container finished" podID="b9b792fe-ab24-4587-a1f6-2ad194c9eb83" containerID="e84fbc857ef08fe10ca354b61edee337ef79228822f2db1fad01f24f50ee52b0" exitCode=0 Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.809232 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-vc6g6" event={"ID":"b9b792fe-ab24-4587-a1f6-2ad194c9eb83","Type":"ContainerDied","Data":"e84fbc857ef08fe10ca354b61edee337ef79228822f2db1fad01f24f50ee52b0"} Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.810604 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"2f52a524-225c-4609-8cb0-0ec800e14aad","Type":"ContainerStarted","Data":"08c6fac38a08a5fdb58e625149872355ce37f196678173626a9383f40610965f"} Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.813003 4967 generic.go:334] "Generic (PLEG): container finished" podID="803e1df8-1759-49e3-83c8-abb1abf6fff4" containerID="aa745d25bad83cb7ef4dbc45d79aa4de11c01a0d95b47616a839e040bdee5b6b" exitCode=0 Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.813083 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-5vbgv" event={"ID":"803e1df8-1759-49e3-83c8-abb1abf6fff4","Type":"ContainerDied","Data":"aa745d25bad83cb7ef4dbc45d79aa4de11c01a0d95b47616a839e040bdee5b6b"} Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.825939 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca","Type":"ContainerStarted","Data":"72bd6a339fddaf7fa71e95bfca3e473cddd5af1f51c6eeb32ab63e9265c7d898"} Oct 11 04:06:12 crc kubenswrapper[4967]: I1011 04:06:12.872052 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-ovs-7b5rp"] Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.074763 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lxnrr"] Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.090577 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-675f4bcbfc-lxnrr"] Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.101523 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-8wsqp"] Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.107197 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-78dd6ddcc-8wsqp"] Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.366115 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.370475 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.375915 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovndbcluster-nb-ovndbs" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.376189 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-config" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.376547 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovndbcluster-nb-scripts" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.383231 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovncluster-ovndbcluster-nb-dockercfg-nmz86" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.387723 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.434683 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsmpx\" (UniqueName: \"kubernetes.io/projected/90a0e0c5-70a6-4330-986c-a09d244f781b-kube-api-access-hsmpx\") pod \"ovsdbserver-nb-0\" (UID: \"90a0e0c5-70a6-4330-986c-a09d244f781b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.434722 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90a0e0c5-70a6-4330-986c-a09d244f781b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"90a0e0c5-70a6-4330-986c-a09d244f781b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.434743 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/90a0e0c5-70a6-4330-986c-a09d244f781b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"90a0e0c5-70a6-4330-986c-a09d244f781b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.434808 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/90a0e0c5-70a6-4330-986c-a09d244f781b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"90a0e0c5-70a6-4330-986c-a09d244f781b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.434823 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90a0e0c5-70a6-4330-986c-a09d244f781b-config\") pod \"ovsdbserver-nb-0\" (UID: \"90a0e0c5-70a6-4330-986c-a09d244f781b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.434869 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/90a0e0c5-70a6-4330-986c-a09d244f781b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"90a0e0c5-70a6-4330-986c-a09d244f781b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.434889 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"90a0e0c5-70a6-4330-986c-a09d244f781b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.434903 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/90a0e0c5-70a6-4330-986c-a09d244f781b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"90a0e0c5-70a6-4330-986c-a09d244f781b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.459558 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-metrics-2n27p"] Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.460628 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-2n27p" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.464279 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-2n27p"] Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.471392 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-metrics-config" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.536628 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"90a0e0c5-70a6-4330-986c-a09d244f781b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.536716 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4g2vj\" (UniqueName: \"kubernetes.io/projected/6df16c29-dfa6-4972-9b1c-c1190bb1ca2a-kube-api-access-4g2vj\") pod \"ovn-controller-metrics-2n27p\" (UID: \"6df16c29-dfa6-4972-9b1c-c1190bb1ca2a\") " pod="openstack/ovn-controller-metrics-2n27p" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.536747 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/90a0e0c5-70a6-4330-986c-a09d244f781b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"90a0e0c5-70a6-4330-986c-a09d244f781b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.536785 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df16c29-dfa6-4972-9b1c-c1190bb1ca2a-combined-ca-bundle\") pod \"ovn-controller-metrics-2n27p\" (UID: \"6df16c29-dfa6-4972-9b1c-c1190bb1ca2a\") " pod="openstack/ovn-controller-metrics-2n27p" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.536819 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/90a0e0c5-70a6-4330-986c-a09d244f781b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"90a0e0c5-70a6-4330-986c-a09d244f781b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.536848 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/90a0e0c5-70a6-4330-986c-a09d244f781b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"90a0e0c5-70a6-4330-986c-a09d244f781b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.536878 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsmpx\" (UniqueName: \"kubernetes.io/projected/90a0e0c5-70a6-4330-986c-a09d244f781b-kube-api-access-hsmpx\") pod \"ovsdbserver-nb-0\" (UID: \"90a0e0c5-70a6-4330-986c-a09d244f781b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.536902 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90a0e0c5-70a6-4330-986c-a09d244f781b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"90a0e0c5-70a6-4330-986c-a09d244f781b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.536928 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/90a0e0c5-70a6-4330-986c-a09d244f781b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"90a0e0c5-70a6-4330-986c-a09d244f781b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.536958 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6df16c29-dfa6-4972-9b1c-c1190bb1ca2a-config\") pod \"ovn-controller-metrics-2n27p\" (UID: \"6df16c29-dfa6-4972-9b1c-c1190bb1ca2a\") " pod="openstack/ovn-controller-metrics-2n27p" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.537015 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/6df16c29-dfa6-4972-9b1c-c1190bb1ca2a-ovn-rundir\") pod \"ovn-controller-metrics-2n27p\" (UID: \"6df16c29-dfa6-4972-9b1c-c1190bb1ca2a\") " pod="openstack/ovn-controller-metrics-2n27p" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.537038 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/6df16c29-dfa6-4972-9b1c-c1190bb1ca2a-ovs-rundir\") pod \"ovn-controller-metrics-2n27p\" (UID: \"6df16c29-dfa6-4972-9b1c-c1190bb1ca2a\") " pod="openstack/ovn-controller-metrics-2n27p" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.537063 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90a0e0c5-70a6-4330-986c-a09d244f781b-config\") pod \"ovsdbserver-nb-0\" (UID: \"90a0e0c5-70a6-4330-986c-a09d244f781b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.537105 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6df16c29-dfa6-4972-9b1c-c1190bb1ca2a-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-2n27p\" (UID: \"6df16c29-dfa6-4972-9b1c-c1190bb1ca2a\") " pod="openstack/ovn-controller-metrics-2n27p" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.537503 4967 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"90a0e0c5-70a6-4330-986c-a09d244f781b\") device mount path \"/mnt/openstack/pv07\"" pod="openstack/ovsdbserver-nb-0" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.540186 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdb-rundir\" (UniqueName: \"kubernetes.io/empty-dir/90a0e0c5-70a6-4330-986c-a09d244f781b-ovsdb-rundir\") pod \"ovsdbserver-nb-0\" (UID: \"90a0e0c5-70a6-4330-986c-a09d244f781b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.541825 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/90a0e0c5-70a6-4330-986c-a09d244f781b-config\") pod \"ovsdbserver-nb-0\" (UID: \"90a0e0c5-70a6-4330-986c-a09d244f781b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.541991 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/90a0e0c5-70a6-4330-986c-a09d244f781b-scripts\") pod \"ovsdbserver-nb-0\" (UID: \"90a0e0c5-70a6-4330-986c-a09d244f781b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.546887 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/90a0e0c5-70a6-4330-986c-a09d244f781b-metrics-certs-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"90a0e0c5-70a6-4330-986c-a09d244f781b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.554532 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsmpx\" (UniqueName: \"kubernetes.io/projected/90a0e0c5-70a6-4330-986c-a09d244f781b-kube-api-access-hsmpx\") pod \"ovsdbserver-nb-0\" (UID: \"90a0e0c5-70a6-4330-986c-a09d244f781b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.561312 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/90a0e0c5-70a6-4330-986c-a09d244f781b-combined-ca-bundle\") pod \"ovsdbserver-nb-0\" (UID: \"90a0e0c5-70a6-4330-986c-a09d244f781b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.562889 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb-tls-certs\" (UniqueName: \"kubernetes.io/secret/90a0e0c5-70a6-4330-986c-a09d244f781b-ovsdbserver-nb-tls-certs\") pod \"ovsdbserver-nb-0\" (UID: \"90a0e0c5-70a6-4330-986c-a09d244f781b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.581471 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage07-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage07-crc\") pod \"ovsdbserver-nb-0\" (UID: \"90a0e0c5-70a6-4330-986c-a09d244f781b\") " pod="openstack/ovsdbserver-nb-0" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.606145 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5vbgv"] Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.640438 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/6df16c29-dfa6-4972-9b1c-c1190bb1ca2a-ovn-rundir\") pod \"ovn-controller-metrics-2n27p\" (UID: \"6df16c29-dfa6-4972-9b1c-c1190bb1ca2a\") " pod="openstack/ovn-controller-metrics-2n27p" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.640484 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/6df16c29-dfa6-4972-9b1c-c1190bb1ca2a-ovs-rundir\") pod \"ovn-controller-metrics-2n27p\" (UID: \"6df16c29-dfa6-4972-9b1c-c1190bb1ca2a\") " pod="openstack/ovn-controller-metrics-2n27p" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.640502 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6df16c29-dfa6-4972-9b1c-c1190bb1ca2a-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-2n27p\" (UID: \"6df16c29-dfa6-4972-9b1c-c1190bb1ca2a\") " pod="openstack/ovn-controller-metrics-2n27p" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.640565 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4g2vj\" (UniqueName: \"kubernetes.io/projected/6df16c29-dfa6-4972-9b1c-c1190bb1ca2a-kube-api-access-4g2vj\") pod \"ovn-controller-metrics-2n27p\" (UID: \"6df16c29-dfa6-4972-9b1c-c1190bb1ca2a\") " pod="openstack/ovn-controller-metrics-2n27p" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.640596 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df16c29-dfa6-4972-9b1c-c1190bb1ca2a-combined-ca-bundle\") pod \"ovn-controller-metrics-2n27p\" (UID: \"6df16c29-dfa6-4972-9b1c-c1190bb1ca2a\") " pod="openstack/ovn-controller-metrics-2n27p" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.640637 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6df16c29-dfa6-4972-9b1c-c1190bb1ca2a-config\") pod \"ovn-controller-metrics-2n27p\" (UID: \"6df16c29-dfa6-4972-9b1c-c1190bb1ca2a\") " pod="openstack/ovn-controller-metrics-2n27p" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.641276 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6df16c29-dfa6-4972-9b1c-c1190bb1ca2a-config\") pod \"ovn-controller-metrics-2n27p\" (UID: \"6df16c29-dfa6-4972-9b1c-c1190bb1ca2a\") " pod="openstack/ovn-controller-metrics-2n27p" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.641506 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/host-path/6df16c29-dfa6-4972-9b1c-c1190bb1ca2a-ovn-rundir\") pod \"ovn-controller-metrics-2n27p\" (UID: \"6df16c29-dfa6-4972-9b1c-c1190bb1ca2a\") " pod="openstack/ovn-controller-metrics-2n27p" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.641544 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovs-rundir\" (UniqueName: \"kubernetes.io/host-path/6df16c29-dfa6-4972-9b1c-c1190bb1ca2a-ovs-rundir\") pod \"ovn-controller-metrics-2n27p\" (UID: \"6df16c29-dfa6-4972-9b1c-c1190bb1ca2a\") " pod="openstack/ovn-controller-metrics-2n27p" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.646623 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-7pf4q"] Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.646634 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6df16c29-dfa6-4972-9b1c-c1190bb1ca2a-combined-ca-bundle\") pod \"ovn-controller-metrics-2n27p\" (UID: \"6df16c29-dfa6-4972-9b1c-c1190bb1ca2a\") " pod="openstack/ovn-controller-metrics-2n27p" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.650908 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-7pf4q" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.665433 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-sb" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.665450 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/6df16c29-dfa6-4972-9b1c-c1190bb1ca2a-metrics-certs-tls-certs\") pod \"ovn-controller-metrics-2n27p\" (UID: \"6df16c29-dfa6-4972-9b1c-c1190bb1ca2a\") " pod="openstack/ovn-controller-metrics-2n27p" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.678766 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-7pf4q"] Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.692603 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4g2vj\" (UniqueName: \"kubernetes.io/projected/6df16c29-dfa6-4972-9b1c-c1190bb1ca2a-kube-api-access-4g2vj\") pod \"ovn-controller-metrics-2n27p\" (UID: \"6df16c29-dfa6-4972-9b1c-c1190bb1ca2a\") " pod="openstack/ovn-controller-metrics-2n27p" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.693129 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovsdbserver-nb-0" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.741776 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2bf19d4-bf7a-414a-9ad5-f603c3ef5575-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-7pf4q\" (UID: \"b2bf19d4-bf7a-414a-9ad5-f603c3ef5575\") " pod="openstack/dnsmasq-dns-6bc7876d45-7pf4q" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.741828 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2bf19d4-bf7a-414a-9ad5-f603c3ef5575-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-7pf4q\" (UID: \"b2bf19d4-bf7a-414a-9ad5-f603c3ef5575\") " pod="openstack/dnsmasq-dns-6bc7876d45-7pf4q" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.741993 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2bf19d4-bf7a-414a-9ad5-f603c3ef5575-config\") pod \"dnsmasq-dns-6bc7876d45-7pf4q\" (UID: \"b2bf19d4-bf7a-414a-9ad5-f603c3ef5575\") " pod="openstack/dnsmasq-dns-6bc7876d45-7pf4q" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.742057 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-chkqw\" (UniqueName: \"kubernetes.io/projected/b2bf19d4-bf7a-414a-9ad5-f603c3ef5575-kube-api-access-chkqw\") pod \"dnsmasq-dns-6bc7876d45-7pf4q\" (UID: \"b2bf19d4-bf7a-414a-9ad5-f603c3ef5575\") " pod="openstack/dnsmasq-dns-6bc7876d45-7pf4q" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.799170 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-metrics-2n27p" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.831851 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-5vbgv" event={"ID":"803e1df8-1759-49e3-83c8-abb1abf6fff4","Type":"ContainerStarted","Data":"ae8b6aaddb3c1caf0effc305423450c48d67717170cbfb2b4916cc95ff4fa064"} Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.832858 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-57d769cc4f-5vbgv" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.834780 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7b5rp" event={"ID":"4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20","Type":"ContainerStarted","Data":"fe7663c244ab09031f6cf269801dc8bcbd0c96a6d6cc23a5695b16814805273b"} Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.837288 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-vc6g6" event={"ID":"b9b792fe-ab24-4587-a1f6-2ad194c9eb83","Type":"ContainerStarted","Data":"da23146880d15c1cb30395f5700e1386827eb8f1150aa9c496aa9c35382d0776"} Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.837807 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-5ccc8479f9-vc6g6" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.843098 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2bf19d4-bf7a-414a-9ad5-f603c3ef5575-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-7pf4q\" (UID: \"b2bf19d4-bf7a-414a-9ad5-f603c3ef5575\") " pod="openstack/dnsmasq-dns-6bc7876d45-7pf4q" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.843177 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2bf19d4-bf7a-414a-9ad5-f603c3ef5575-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-7pf4q\" (UID: \"b2bf19d4-bf7a-414a-9ad5-f603c3ef5575\") " pod="openstack/dnsmasq-dns-6bc7876d45-7pf4q" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.843242 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2bf19d4-bf7a-414a-9ad5-f603c3ef5575-config\") pod \"dnsmasq-dns-6bc7876d45-7pf4q\" (UID: \"b2bf19d4-bf7a-414a-9ad5-f603c3ef5575\") " pod="openstack/dnsmasq-dns-6bc7876d45-7pf4q" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.843273 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-chkqw\" (UniqueName: \"kubernetes.io/projected/b2bf19d4-bf7a-414a-9ad5-f603c3ef5575-kube-api-access-chkqw\") pod \"dnsmasq-dns-6bc7876d45-7pf4q\" (UID: \"b2bf19d4-bf7a-414a-9ad5-f603c3ef5575\") " pod="openstack/dnsmasq-dns-6bc7876d45-7pf4q" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.844354 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2bf19d4-bf7a-414a-9ad5-f603c3ef5575-dns-svc\") pod \"dnsmasq-dns-6bc7876d45-7pf4q\" (UID: \"b2bf19d4-bf7a-414a-9ad5-f603c3ef5575\") " pod="openstack/dnsmasq-dns-6bc7876d45-7pf4q" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.844411 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2bf19d4-bf7a-414a-9ad5-f603c3ef5575-ovsdbserver-sb\") pod \"dnsmasq-dns-6bc7876d45-7pf4q\" (UID: \"b2bf19d4-bf7a-414a-9ad5-f603c3ef5575\") " pod="openstack/dnsmasq-dns-6bc7876d45-7pf4q" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.844423 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2bf19d4-bf7a-414a-9ad5-f603c3ef5575-config\") pod \"dnsmasq-dns-6bc7876d45-7pf4q\" (UID: \"b2bf19d4-bf7a-414a-9ad5-f603c3ef5575\") " pod="openstack/dnsmasq-dns-6bc7876d45-7pf4q" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.858840 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-57d769cc4f-5vbgv" podStartSLOduration=14.401434742 podStartE2EDuration="14.858795221s" podCreationTimestamp="2025-10-11 04:05:59 +0000 UTC" firstStartedPulling="2025-10-11 04:06:11.525571519 +0000 UTC m=+899.488780462" lastFinishedPulling="2025-10-11 04:06:11.982932008 +0000 UTC m=+899.946140941" observedRunningTime="2025-10-11 04:06:13.852989695 +0000 UTC m=+901.816198628" watchObservedRunningTime="2025-10-11 04:06:13.858795221 +0000 UTC m=+901.822004154" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.864726 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-chkqw\" (UniqueName: \"kubernetes.io/projected/b2bf19d4-bf7a-414a-9ad5-f603c3ef5575-kube-api-access-chkqw\") pod \"dnsmasq-dns-6bc7876d45-7pf4q\" (UID: \"b2bf19d4-bf7a-414a-9ad5-f603c3ef5575\") " pod="openstack/dnsmasq-dns-6bc7876d45-7pf4q" Oct 11 04:06:13 crc kubenswrapper[4967]: I1011 04:06:13.875403 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-5ccc8479f9-vc6g6" podStartSLOduration=6.4750585449999996 podStartE2EDuration="14.875377173s" podCreationTimestamp="2025-10-11 04:05:59 +0000 UTC" firstStartedPulling="2025-10-11 04:06:03.30223128 +0000 UTC m=+891.265440223" lastFinishedPulling="2025-10-11 04:06:11.702549918 +0000 UTC m=+899.665758851" observedRunningTime="2025-10-11 04:06:13.868246675 +0000 UTC m=+901.831455608" watchObservedRunningTime="2025-10-11 04:06:13.875377173 +0000 UTC m=+901.838586106" Oct 11 04:06:14 crc kubenswrapper[4967]: I1011 04:06:14.045887 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-7pf4q" Oct 11 04:06:14 crc kubenswrapper[4967]: I1011 04:06:14.090290 4967 patch_prober.go:28] interesting pod/machine-config-daemon-sgrv2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:06:14 crc kubenswrapper[4967]: I1011 04:06:14.090357 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:06:14 crc kubenswrapper[4967]: I1011 04:06:14.825565 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="854c2370-c43a-4c50-99ab-5d935456c2ae" path="/var/lib/kubelet/pods/854c2370-c43a-4c50-99ab-5d935456c2ae/volumes" Oct 11 04:06:14 crc kubenswrapper[4967]: I1011 04:06:14.826349 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f980d5dd-67f3-4714-a626-d521efde4916" path="/var/lib/kubelet/pods/f980d5dd-67f3-4714-a626-d521efde4916/volumes" Oct 11 04:06:14 crc kubenswrapper[4967]: I1011 04:06:14.845160 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-57d769cc4f-5vbgv" podUID="803e1df8-1759-49e3-83c8-abb1abf6fff4" containerName="dnsmasq-dns" containerID="cri-o://ae8b6aaddb3c1caf0effc305423450c48d67717170cbfb2b4916cc95ff4fa064" gracePeriod=10 Oct 11 04:06:15 crc kubenswrapper[4967]: I1011 04:06:15.862228 4967 generic.go:334] "Generic (PLEG): container finished" podID="803e1df8-1759-49e3-83c8-abb1abf6fff4" containerID="ae8b6aaddb3c1caf0effc305423450c48d67717170cbfb2b4916cc95ff4fa064" exitCode=0 Oct 11 04:06:15 crc kubenswrapper[4967]: I1011 04:06:15.862342 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-5vbgv" event={"ID":"803e1df8-1759-49e3-83c8-abb1abf6fff4","Type":"ContainerDied","Data":"ae8b6aaddb3c1caf0effc305423450c48d67717170cbfb2b4916cc95ff4fa064"} Oct 11 04:06:19 crc kubenswrapper[4967]: I1011 04:06:19.788246 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-5ccc8479f9-vc6g6" Oct 11 04:06:21 crc kubenswrapper[4967]: I1011 04:06:21.755098 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-4txc5"] Oct 11 04:06:21 crc kubenswrapper[4967]: I1011 04:06:21.757458 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4txc5" Oct 11 04:06:21 crc kubenswrapper[4967]: I1011 04:06:21.766124 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4txc5"] Oct 11 04:06:21 crc kubenswrapper[4967]: I1011 04:06:21.846192 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-5vbgv" Oct 11 04:06:21 crc kubenswrapper[4967]: I1011 04:06:21.881897 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/245cf6c7-f572-4395-afdf-ef0795ff6c05-catalog-content\") pod \"certified-operators-4txc5\" (UID: \"245cf6c7-f572-4395-afdf-ef0795ff6c05\") " pod="openshift-marketplace/certified-operators-4txc5" Oct 11 04:06:21 crc kubenswrapper[4967]: I1011 04:06:21.882094 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/245cf6c7-f572-4395-afdf-ef0795ff6c05-utilities\") pod \"certified-operators-4txc5\" (UID: \"245cf6c7-f572-4395-afdf-ef0795ff6c05\") " pod="openshift-marketplace/certified-operators-4txc5" Oct 11 04:06:21 crc kubenswrapper[4967]: I1011 04:06:21.882157 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-swjq2\" (UniqueName: \"kubernetes.io/projected/245cf6c7-f572-4395-afdf-ef0795ff6c05-kube-api-access-swjq2\") pod \"certified-operators-4txc5\" (UID: \"245cf6c7-f572-4395-afdf-ef0795ff6c05\") " pod="openshift-marketplace/certified-operators-4txc5" Oct 11 04:06:21 crc kubenswrapper[4967]: I1011 04:06:21.908833 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-57d769cc4f-5vbgv" event={"ID":"803e1df8-1759-49e3-83c8-abb1abf6fff4","Type":"ContainerDied","Data":"b2f782f032c59630f6f26294695df03be2b32b713cb7a24fc687666e9fc9062c"} Oct 11 04:06:21 crc kubenswrapper[4967]: I1011 04:06:21.908884 4967 scope.go:117] "RemoveContainer" containerID="ae8b6aaddb3c1caf0effc305423450c48d67717170cbfb2b4916cc95ff4fa064" Oct 11 04:06:21 crc kubenswrapper[4967]: I1011 04:06:21.908916 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-57d769cc4f-5vbgv" Oct 11 04:06:21 crc kubenswrapper[4967]: I1011 04:06:21.983398 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f66tt\" (UniqueName: \"kubernetes.io/projected/803e1df8-1759-49e3-83c8-abb1abf6fff4-kube-api-access-f66tt\") pod \"803e1df8-1759-49e3-83c8-abb1abf6fff4\" (UID: \"803e1df8-1759-49e3-83c8-abb1abf6fff4\") " Oct 11 04:06:21 crc kubenswrapper[4967]: I1011 04:06:21.983592 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/803e1df8-1759-49e3-83c8-abb1abf6fff4-config\") pod \"803e1df8-1759-49e3-83c8-abb1abf6fff4\" (UID: \"803e1df8-1759-49e3-83c8-abb1abf6fff4\") " Oct 11 04:06:21 crc kubenswrapper[4967]: I1011 04:06:21.983751 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/803e1df8-1759-49e3-83c8-abb1abf6fff4-dns-svc\") pod \"803e1df8-1759-49e3-83c8-abb1abf6fff4\" (UID: \"803e1df8-1759-49e3-83c8-abb1abf6fff4\") " Oct 11 04:06:21 crc kubenswrapper[4967]: I1011 04:06:21.984068 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/245cf6c7-f572-4395-afdf-ef0795ff6c05-catalog-content\") pod \"certified-operators-4txc5\" (UID: \"245cf6c7-f572-4395-afdf-ef0795ff6c05\") " pod="openshift-marketplace/certified-operators-4txc5" Oct 11 04:06:21 crc kubenswrapper[4967]: I1011 04:06:21.984211 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/245cf6c7-f572-4395-afdf-ef0795ff6c05-utilities\") pod \"certified-operators-4txc5\" (UID: \"245cf6c7-f572-4395-afdf-ef0795ff6c05\") " pod="openshift-marketplace/certified-operators-4txc5" Oct 11 04:06:21 crc kubenswrapper[4967]: I1011 04:06:21.984271 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-swjq2\" (UniqueName: \"kubernetes.io/projected/245cf6c7-f572-4395-afdf-ef0795ff6c05-kube-api-access-swjq2\") pod \"certified-operators-4txc5\" (UID: \"245cf6c7-f572-4395-afdf-ef0795ff6c05\") " pod="openshift-marketplace/certified-operators-4txc5" Oct 11 04:06:21 crc kubenswrapper[4967]: I1011 04:06:21.984524 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/245cf6c7-f572-4395-afdf-ef0795ff6c05-catalog-content\") pod \"certified-operators-4txc5\" (UID: \"245cf6c7-f572-4395-afdf-ef0795ff6c05\") " pod="openshift-marketplace/certified-operators-4txc5" Oct 11 04:06:21 crc kubenswrapper[4967]: I1011 04:06:21.984769 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/245cf6c7-f572-4395-afdf-ef0795ff6c05-utilities\") pod \"certified-operators-4txc5\" (UID: \"245cf6c7-f572-4395-afdf-ef0795ff6c05\") " pod="openshift-marketplace/certified-operators-4txc5" Oct 11 04:06:22 crc kubenswrapper[4967]: I1011 04:06:21.998505 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/803e1df8-1759-49e3-83c8-abb1abf6fff4-kube-api-access-f66tt" (OuterVolumeSpecName: "kube-api-access-f66tt") pod "803e1df8-1759-49e3-83c8-abb1abf6fff4" (UID: "803e1df8-1759-49e3-83c8-abb1abf6fff4"). InnerVolumeSpecName "kube-api-access-f66tt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:06:22 crc kubenswrapper[4967]: I1011 04:06:22.002203 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-swjq2\" (UniqueName: \"kubernetes.io/projected/245cf6c7-f572-4395-afdf-ef0795ff6c05-kube-api-access-swjq2\") pod \"certified-operators-4txc5\" (UID: \"245cf6c7-f572-4395-afdf-ef0795ff6c05\") " pod="openshift-marketplace/certified-operators-4txc5" Oct 11 04:06:22 crc kubenswrapper[4967]: I1011 04:06:22.023935 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/803e1df8-1759-49e3-83c8-abb1abf6fff4-config" (OuterVolumeSpecName: "config") pod "803e1df8-1759-49e3-83c8-abb1abf6fff4" (UID: "803e1df8-1759-49e3-83c8-abb1abf6fff4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:06:22 crc kubenswrapper[4967]: I1011 04:06:22.037115 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/803e1df8-1759-49e3-83c8-abb1abf6fff4-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "803e1df8-1759-49e3-83c8-abb1abf6fff4" (UID: "803e1df8-1759-49e3-83c8-abb1abf6fff4"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:06:22 crc kubenswrapper[4967]: I1011 04:06:22.085513 4967 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/803e1df8-1759-49e3-83c8-abb1abf6fff4-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:22 crc kubenswrapper[4967]: I1011 04:06:22.085542 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f66tt\" (UniqueName: \"kubernetes.io/projected/803e1df8-1759-49e3-83c8-abb1abf6fff4-kube-api-access-f66tt\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:22 crc kubenswrapper[4967]: I1011 04:06:22.085553 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/803e1df8-1759-49e3-83c8-abb1abf6fff4-config\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:22 crc kubenswrapper[4967]: I1011 04:06:22.143931 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4txc5" Oct 11 04:06:22 crc kubenswrapper[4967]: I1011 04:06:22.241040 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5vbgv"] Oct 11 04:06:22 crc kubenswrapper[4967]: I1011 04:06:22.251989 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-57d769cc4f-5vbgv"] Oct 11 04:06:22 crc kubenswrapper[4967]: I1011 04:06:22.745480 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-7pf4q"] Oct 11 04:06:22 crc kubenswrapper[4967]: I1011 04:06:22.830461 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="803e1df8-1759-49e3-83c8-abb1abf6fff4" path="/var/lib/kubelet/pods/803e1df8-1759-49e3-83c8-abb1abf6fff4/volumes" Oct 11 04:06:22 crc kubenswrapper[4967]: I1011 04:06:22.832670 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-metrics-2n27p"] Oct 11 04:06:22 crc kubenswrapper[4967]: I1011 04:06:22.884342 4967 scope.go:117] "RemoveContainer" containerID="aa745d25bad83cb7ef4dbc45d79aa4de11c01a0d95b47616a839e040bdee5b6b" Oct 11 04:06:22 crc kubenswrapper[4967]: W1011 04:06:22.891817 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6df16c29_dfa6_4972_9b1c_c1190bb1ca2a.slice/crio-b65be87981fac4e88eb11127163feba9797d92a0b3b47c6a3a9f002adf85df2d WatchSource:0}: Error finding container b65be87981fac4e88eb11127163feba9797d92a0b3b47c6a3a9f002adf85df2d: Status 404 returned error can't find the container with id b65be87981fac4e88eb11127163feba9797d92a0b3b47c6a3a9f002adf85df2d Oct 11 04:06:22 crc kubenswrapper[4967]: I1011 04:06:22.921124 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-7pf4q" event={"ID":"b2bf19d4-bf7a-414a-9ad5-f603c3ef5575","Type":"ContainerStarted","Data":"a6d7649366db251a3802e36b053c218fbbc665bce4d593ba52ca6cb332139845"} Oct 11 04:06:22 crc kubenswrapper[4967]: I1011 04:06:22.922384 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-2n27p" event={"ID":"6df16c29-dfa6-4972-9b1c-c1190bb1ca2a","Type":"ContainerStarted","Data":"b65be87981fac4e88eb11127163feba9797d92a0b3b47c6a3a9f002adf85df2d"} Oct 11 04:06:23 crc kubenswrapper[4967]: I1011 04:06:23.270862 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovsdbserver-nb-0"] Oct 11 04:06:23 crc kubenswrapper[4967]: I1011 04:06:23.459989 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-4txc5"] Oct 11 04:06:23 crc kubenswrapper[4967]: W1011 04:06:23.483109 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod90a0e0c5_70a6_4330_986c_a09d244f781b.slice/crio-beb03a21dcff0a8d6da7d8f0e24796f37dca1c5741e417c66cc34addc657447e WatchSource:0}: Error finding container beb03a21dcff0a8d6da7d8f0e24796f37dca1c5741e417c66cc34addc657447e: Status 404 returned error can't find the container with id beb03a21dcff0a8d6da7d8f0e24796f37dca1c5741e417c66cc34addc657447e Oct 11 04:06:23 crc kubenswrapper[4967]: W1011 04:06:23.486154 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod245cf6c7_f572_4395_afdf_ef0795ff6c05.slice/crio-f9f0de980ee68fa7e0fba63fccc20dedba81145da333b2de684c4d5b9246542c WatchSource:0}: Error finding container f9f0de980ee68fa7e0fba63fccc20dedba81145da333b2de684c4d5b9246542c: Status 404 returned error can't find the container with id f9f0de980ee68fa7e0fba63fccc20dedba81145da333b2de684c4d5b9246542c Oct 11 04:06:23 crc kubenswrapper[4967]: I1011 04:06:23.933950 4967 generic.go:334] "Generic (PLEG): container finished" podID="245cf6c7-f572-4395-afdf-ef0795ff6c05" containerID="e3856541bbe583a18caf3cc5114975737812605d997212f79240d0d7c09ca8fa" exitCode=0 Oct 11 04:06:23 crc kubenswrapper[4967]: I1011 04:06:23.934087 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4txc5" event={"ID":"245cf6c7-f572-4395-afdf-ef0795ff6c05","Type":"ContainerDied","Data":"e3856541bbe583a18caf3cc5114975737812605d997212f79240d0d7c09ca8fa"} Oct 11 04:06:23 crc kubenswrapper[4967]: I1011 04:06:23.934626 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4txc5" event={"ID":"245cf6c7-f572-4395-afdf-ef0795ff6c05","Type":"ContainerStarted","Data":"f9f0de980ee68fa7e0fba63fccc20dedba81145da333b2de684c4d5b9246542c"} Oct 11 04:06:23 crc kubenswrapper[4967]: I1011 04:06:23.941645 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/memcached-0" event={"ID":"10ea033f-281c-4519-8da7-93e598f08b11","Type":"ContainerStarted","Data":"26fc0e74ada7fe90cb2ab068b3e63fa345f1250b489a877f7b4fed4e1041a4cb"} Oct 11 04:06:23 crc kubenswrapper[4967]: I1011 04:06:23.941824 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/memcached-0" Oct 11 04:06:23 crc kubenswrapper[4967]: I1011 04:06:23.943715 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ddc18aa7-8599-40c5-a019-972c13162f83","Type":"ContainerStarted","Data":"3cc5e377834d69415ac8bbe0c98e2d842dc2ad9f8378bd0b1ffe73aa8727b8f3"} Oct 11 04:06:23 crc kubenswrapper[4967]: I1011 04:06:23.946055 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"90a0e0c5-70a6-4330-986c-a09d244f781b","Type":"ContainerStarted","Data":"beb03a21dcff0a8d6da7d8f0e24796f37dca1c5741e417c66cc34addc657447e"} Oct 11 04:06:23 crc kubenswrapper[4967]: I1011 04:06:23.947842 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"2f52a524-225c-4609-8cb0-0ec800e14aad","Type":"ContainerStarted","Data":"976004492a85701ed3f52fa7bc8e5867255ef89dbdafc06b13470404e5d20715"} Oct 11 04:06:23 crc kubenswrapper[4967]: I1011 04:06:23.956027 4967 generic.go:334] "Generic (PLEG): container finished" podID="b2bf19d4-bf7a-414a-9ad5-f603c3ef5575" containerID="16622120bb6d6340098cb0f262d55c2ced0e00e7f424b93dc1df8f4e5d074fb8" exitCode=0 Oct 11 04:06:23 crc kubenswrapper[4967]: I1011 04:06:23.956129 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-7pf4q" event={"ID":"b2bf19d4-bf7a-414a-9ad5-f603c3ef5575","Type":"ContainerDied","Data":"16622120bb6d6340098cb0f262d55c2ced0e00e7f424b93dc1df8f4e5d074fb8"} Oct 11 04:06:24 crc kubenswrapper[4967]: I1011 04:06:24.036180 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/memcached-0" podStartSLOduration=9.822104315 podStartE2EDuration="20.036159339s" podCreationTimestamp="2025-10-11 04:06:04 +0000 UTC" firstStartedPulling="2025-10-11 04:06:12.390556193 +0000 UTC m=+900.353765126" lastFinishedPulling="2025-10-11 04:06:22.604611217 +0000 UTC m=+910.567820150" observedRunningTime="2025-10-11 04:06:24.017353765 +0000 UTC m=+911.980562698" watchObservedRunningTime="2025-10-11 04:06:24.036159339 +0000 UTC m=+911.999368272" Oct 11 04:06:24 crc kubenswrapper[4967]: I1011 04:06:24.967358 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3d923c19-a45c-4cb7-9f19-e770c5fe66ed","Type":"ContainerStarted","Data":"90a1eab201a7f4eef8d6e63d4754f3e37c2c7b5857764dda95a5217b63654438"} Oct 11 04:06:24 crc kubenswrapper[4967]: I1011 04:06:24.982137 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3d120d8c-893d-49d9-8649-453997725ff2","Type":"ContainerStarted","Data":"4504fe990984cab7895c8f853e17d85cd0e0782bbe982534f0c7080a6b839521"} Oct 11 04:06:24 crc kubenswrapper[4967]: I1011 04:06:24.982174 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 11 04:06:24 crc kubenswrapper[4967]: I1011 04:06:24.983953 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-wvrnq" event={"ID":"acad8def-db3f-40a3-a95d-28b7503d4aae","Type":"ContainerStarted","Data":"247fe2e9e0fa998e62e867a52bda77885d017e7f572e5ebf31a45b751f3e6d33"} Oct 11 04:06:24 crc kubenswrapper[4967]: I1011 04:06:24.984333 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-wvrnq" Oct 11 04:06:24 crc kubenswrapper[4967]: I1011 04:06:24.988419 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-7pf4q" event={"ID":"b2bf19d4-bf7a-414a-9ad5-f603c3ef5575","Type":"ContainerStarted","Data":"209853d07705f6072700ccd7046271b186718ad2edbda7dff16d31ad8230b8fa"} Oct 11 04:06:24 crc kubenswrapper[4967]: I1011 04:06:24.988891 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6bc7876d45-7pf4q" Oct 11 04:06:24 crc kubenswrapper[4967]: I1011 04:06:24.997410 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca","Type":"ContainerStarted","Data":"63205383c92f92a8c18ce0e41228fbb364c436076bfae05d05051d90e1bc2c0e"} Oct 11 04:06:24 crc kubenswrapper[4967]: I1011 04:06:24.999772 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7b5rp" event={"ID":"4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20","Type":"ContainerStarted","Data":"13fade01ea20acb63a360d3d8771f77778f68bbd7a854e80c24d6fb9fc8f6940"} Oct 11 04:06:25 crc kubenswrapper[4967]: I1011 04:06:25.004381 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=8.114471419 podStartE2EDuration="19.00436943s" podCreationTimestamp="2025-10-11 04:06:06 +0000 UTC" firstStartedPulling="2025-10-11 04:06:12.734340681 +0000 UTC m=+900.697549614" lastFinishedPulling="2025-10-11 04:06:23.624238692 +0000 UTC m=+911.587447625" observedRunningTime="2025-10-11 04:06:25.003592082 +0000 UTC m=+912.966801015" watchObservedRunningTime="2025-10-11 04:06:25.00436943 +0000 UTC m=+912.967578363" Oct 11 04:06:25 crc kubenswrapper[4967]: I1011 04:06:25.023056 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-wvrnq" podStartSLOduration=4.609310931 podStartE2EDuration="15.02301472s" podCreationTimestamp="2025-10-11 04:06:10 +0000 UTC" firstStartedPulling="2025-10-11 04:06:12.658192542 +0000 UTC m=+900.621401475" lastFinishedPulling="2025-10-11 04:06:23.071896331 +0000 UTC m=+911.035105264" observedRunningTime="2025-10-11 04:06:25.022009797 +0000 UTC m=+912.985218730" watchObservedRunningTime="2025-10-11 04:06:25.02301472 +0000 UTC m=+912.986223653" Oct 11 04:06:25 crc kubenswrapper[4967]: I1011 04:06:25.046862 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6bc7876d45-7pf4q" podStartSLOduration=12.046845503 podStartE2EDuration="12.046845503s" podCreationTimestamp="2025-10-11 04:06:13 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:06:25.042024929 +0000 UTC m=+913.005233892" watchObservedRunningTime="2025-10-11 04:06:25.046845503 +0000 UTC m=+913.010054426" Oct 11 04:06:25 crc kubenswrapper[4967]: I1011 04:06:25.095996 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-57d769cc4f-5vbgv" podUID="803e1df8-1759-49e3-83c8-abb1abf6fff4" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.98:5353: i/o timeout" Oct 11 04:06:26 crc kubenswrapper[4967]: I1011 04:06:26.006186 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0c315653-59d1-429f-be9c-c1ac826cff49","Type":"ContainerStarted","Data":"55e8d168f569f266bd42f0f73e79f782f1b67630b934f6721fb999704f507332"} Oct 11 04:06:26 crc kubenswrapper[4967]: I1011 04:06:26.007583 4967 generic.go:334] "Generic (PLEG): container finished" podID="4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20" containerID="13fade01ea20acb63a360d3d8771f77778f68bbd7a854e80c24d6fb9fc8f6940" exitCode=0 Oct 11 04:06:26 crc kubenswrapper[4967]: I1011 04:06:26.007625 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7b5rp" event={"ID":"4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20","Type":"ContainerDied","Data":"13fade01ea20acb63a360d3d8771f77778f68bbd7a854e80c24d6fb9fc8f6940"} Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.020967 4967 generic.go:334] "Generic (PLEG): container finished" podID="245cf6c7-f572-4395-afdf-ef0795ff6c05" containerID="d1f8bdb850efc1c71549c5625047537b78b2be4905fbdb7b18043592d1dbfd6f" exitCode=0 Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.021095 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4txc5" event={"ID":"245cf6c7-f572-4395-afdf-ef0795ff6c05","Type":"ContainerDied","Data":"d1f8bdb850efc1c71549c5625047537b78b2be4905fbdb7b18043592d1dbfd6f"} Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.028500 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7b5rp" event={"ID":"4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20","Type":"ContainerStarted","Data":"13b3626cea6f7938495dce05fb89453446765655117918f921f190b39c41613b"} Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.028547 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-ovs-7b5rp" event={"ID":"4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20","Type":"ContainerStarted","Data":"fcfa4a630b56ab5f946e5580a5e017703aa3a079e69588a194338234f0e8e5e9"} Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.028570 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-7b5rp" Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.028586 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-controller-ovs-7b5rp" Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.038065 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"90a0e0c5-70a6-4330-986c-a09d244f781b","Type":"ContainerStarted","Data":"557293ef8d426ccaf4b99de081a607089e1ef37ec617c914f1fe29a94d8020ef"} Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.038120 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-nb-0" event={"ID":"90a0e0c5-70a6-4330-986c-a09d244f781b","Type":"ContainerStarted","Data":"2b9133cd54aaf6e33e6b462ebea6cef0cbf06ca4fb474b331e207430b3659043"} Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.044223 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-metrics-2n27p" event={"ID":"6df16c29-dfa6-4972-9b1c-c1190bb1ca2a","Type":"ContainerStarted","Data":"cba4ab45b6bcbbe89beb38f0772d5074d49abe45825e1069a38e3139ab29186d"} Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.047395 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovsdbserver-sb-0" event={"ID":"36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca","Type":"ContainerStarted","Data":"3c2c0d22c5ea99601890f5c0aeca6143a2663dc450aa70e5793c154820cd9eb6"} Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.093774 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-sb-0" podStartSLOduration=4.55130655 podStartE2EDuration="18.093756854s" podCreationTimestamp="2025-10-11 04:06:09 +0000 UTC" firstStartedPulling="2025-10-11 04:06:12.764980494 +0000 UTC m=+900.728189427" lastFinishedPulling="2025-10-11 04:06:26.307430798 +0000 UTC m=+914.270639731" observedRunningTime="2025-10-11 04:06:27.090406106 +0000 UTC m=+915.053615079" watchObservedRunningTime="2025-10-11 04:06:27.093756854 +0000 UTC m=+915.056965787" Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.122939 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovsdbserver-nb-0" podStartSLOduration=12.445412562 podStartE2EDuration="15.122918213s" podCreationTimestamp="2025-10-11 04:06:12 +0000 UTC" firstStartedPulling="2025-10-11 04:06:23.487203387 +0000 UTC m=+911.450412320" lastFinishedPulling="2025-10-11 04:06:26.164709038 +0000 UTC m=+914.127917971" observedRunningTime="2025-10-11 04:06:27.113391148 +0000 UTC m=+915.076600091" watchObservedRunningTime="2025-10-11 04:06:27.122918213 +0000 UTC m=+915.086127146" Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.141657 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-metrics-2n27p" podStartSLOduration=10.803446404 podStartE2EDuration="14.141635396s" podCreationTimestamp="2025-10-11 04:06:13 +0000 UTC" firstStartedPulling="2025-10-11 04:06:22.893260783 +0000 UTC m=+910.856469736" lastFinishedPulling="2025-10-11 04:06:26.231449795 +0000 UTC m=+914.194658728" observedRunningTime="2025-10-11 04:06:27.139986146 +0000 UTC m=+915.103195079" watchObservedRunningTime="2025-10-11 04:06:27.141635396 +0000 UTC m=+915.104844329" Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.184675 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-controller-ovs-7b5rp" podStartSLOduration=7.038451548 podStartE2EDuration="17.184659291s" podCreationTimestamp="2025-10-11 04:06:10 +0000 UTC" firstStartedPulling="2025-10-11 04:06:12.912533908 +0000 UTC m=+900.875742851" lastFinishedPulling="2025-10-11 04:06:23.058741651 +0000 UTC m=+911.021950594" observedRunningTime="2025-10-11 04:06:27.181602909 +0000 UTC m=+915.144811842" watchObservedRunningTime="2025-10-11 04:06:27.184659291 +0000 UTC m=+915.147868224" Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.490933 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-7pf4q"] Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.521650 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-8554648995-56sgg"] Oct 11 04:06:27 crc kubenswrapper[4967]: E1011 04:06:27.522057 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="803e1df8-1759-49e3-83c8-abb1abf6fff4" containerName="init" Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.522100 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="803e1df8-1759-49e3-83c8-abb1abf6fff4" containerName="init" Oct 11 04:06:27 crc kubenswrapper[4967]: E1011 04:06:27.522118 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="803e1df8-1759-49e3-83c8-abb1abf6fff4" containerName="dnsmasq-dns" Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.522126 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="803e1df8-1759-49e3-83c8-abb1abf6fff4" containerName="dnsmasq-dns" Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.522323 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="803e1df8-1759-49e3-83c8-abb1abf6fff4" containerName="dnsmasq-dns" Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.524768 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-56sgg" Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.526381 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovsdbserver-nb" Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.533124 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-56sgg"] Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.689042 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc388f4e-6c72-49ba-a4d4-fb7a01e13681-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-56sgg\" (UID: \"bc388f4e-6c72-49ba-a4d4-fb7a01e13681\") " pod="openstack/dnsmasq-dns-8554648995-56sgg" Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.689151 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc388f4e-6c72-49ba-a4d4-fb7a01e13681-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-56sgg\" (UID: \"bc388f4e-6c72-49ba-a4d4-fb7a01e13681\") " pod="openstack/dnsmasq-dns-8554648995-56sgg" Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.689248 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc388f4e-6c72-49ba-a4d4-fb7a01e13681-dns-svc\") pod \"dnsmasq-dns-8554648995-56sgg\" (UID: \"bc388f4e-6c72-49ba-a4d4-fb7a01e13681\") " pod="openstack/dnsmasq-dns-8554648995-56sgg" Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.689291 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc388f4e-6c72-49ba-a4d4-fb7a01e13681-config\") pod \"dnsmasq-dns-8554648995-56sgg\" (UID: \"bc388f4e-6c72-49ba-a4d4-fb7a01e13681\") " pod="openstack/dnsmasq-dns-8554648995-56sgg" Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.689386 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svft6\" (UniqueName: \"kubernetes.io/projected/bc388f4e-6c72-49ba-a4d4-fb7a01e13681-kube-api-access-svft6\") pod \"dnsmasq-dns-8554648995-56sgg\" (UID: \"bc388f4e-6c72-49ba-a4d4-fb7a01e13681\") " pod="openstack/dnsmasq-dns-8554648995-56sgg" Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.790833 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc388f4e-6c72-49ba-a4d4-fb7a01e13681-dns-svc\") pod \"dnsmasq-dns-8554648995-56sgg\" (UID: \"bc388f4e-6c72-49ba-a4d4-fb7a01e13681\") " pod="openstack/dnsmasq-dns-8554648995-56sgg" Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.790872 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc388f4e-6c72-49ba-a4d4-fb7a01e13681-config\") pod \"dnsmasq-dns-8554648995-56sgg\" (UID: \"bc388f4e-6c72-49ba-a4d4-fb7a01e13681\") " pod="openstack/dnsmasq-dns-8554648995-56sgg" Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.790922 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svft6\" (UniqueName: \"kubernetes.io/projected/bc388f4e-6c72-49ba-a4d4-fb7a01e13681-kube-api-access-svft6\") pod \"dnsmasq-dns-8554648995-56sgg\" (UID: \"bc388f4e-6c72-49ba-a4d4-fb7a01e13681\") " pod="openstack/dnsmasq-dns-8554648995-56sgg" Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.790957 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc388f4e-6c72-49ba-a4d4-fb7a01e13681-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-56sgg\" (UID: \"bc388f4e-6c72-49ba-a4d4-fb7a01e13681\") " pod="openstack/dnsmasq-dns-8554648995-56sgg" Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.791002 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc388f4e-6c72-49ba-a4d4-fb7a01e13681-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-56sgg\" (UID: \"bc388f4e-6c72-49ba-a4d4-fb7a01e13681\") " pod="openstack/dnsmasq-dns-8554648995-56sgg" Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.791853 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc388f4e-6c72-49ba-a4d4-fb7a01e13681-config\") pod \"dnsmasq-dns-8554648995-56sgg\" (UID: \"bc388f4e-6c72-49ba-a4d4-fb7a01e13681\") " pod="openstack/dnsmasq-dns-8554648995-56sgg" Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.791852 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc388f4e-6c72-49ba-a4d4-fb7a01e13681-dns-svc\") pod \"dnsmasq-dns-8554648995-56sgg\" (UID: \"bc388f4e-6c72-49ba-a4d4-fb7a01e13681\") " pod="openstack/dnsmasq-dns-8554648995-56sgg" Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.792029 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc388f4e-6c72-49ba-a4d4-fb7a01e13681-ovsdbserver-sb\") pod \"dnsmasq-dns-8554648995-56sgg\" (UID: \"bc388f4e-6c72-49ba-a4d4-fb7a01e13681\") " pod="openstack/dnsmasq-dns-8554648995-56sgg" Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.792104 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc388f4e-6c72-49ba-a4d4-fb7a01e13681-ovsdbserver-nb\") pod \"dnsmasq-dns-8554648995-56sgg\" (UID: \"bc388f4e-6c72-49ba-a4d4-fb7a01e13681\") " pod="openstack/dnsmasq-dns-8554648995-56sgg" Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.814812 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svft6\" (UniqueName: \"kubernetes.io/projected/bc388f4e-6c72-49ba-a4d4-fb7a01e13681-kube-api-access-svft6\") pod \"dnsmasq-dns-8554648995-56sgg\" (UID: \"bc388f4e-6c72-49ba-a4d4-fb7a01e13681\") " pod="openstack/dnsmasq-dns-8554648995-56sgg" Oct 11 04:06:27 crc kubenswrapper[4967]: I1011 04:06:27.874399 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-56sgg" Oct 11 04:06:28 crc kubenswrapper[4967]: I1011 04:06:28.065564 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4txc5" event={"ID":"245cf6c7-f572-4395-afdf-ef0795ff6c05","Type":"ContainerStarted","Data":"f7b0a3c9e42fbbfdff1690f885565d0619d4e490fd7ccfe45a21c3f0eff56fda"} Oct 11 04:06:28 crc kubenswrapper[4967]: I1011 04:06:28.067731 4967 generic.go:334] "Generic (PLEG): container finished" podID="ddc18aa7-8599-40c5-a019-972c13162f83" containerID="3cc5e377834d69415ac8bbe0c98e2d842dc2ad9f8378bd0b1ffe73aa8727b8f3" exitCode=0 Oct 11 04:06:28 crc kubenswrapper[4967]: I1011 04:06:28.067825 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ddc18aa7-8599-40c5-a019-972c13162f83","Type":"ContainerDied","Data":"3cc5e377834d69415ac8bbe0c98e2d842dc2ad9f8378bd0b1ffe73aa8727b8f3"} Oct 11 04:06:28 crc kubenswrapper[4967]: I1011 04:06:28.070686 4967 generic.go:334] "Generic (PLEG): container finished" podID="2f52a524-225c-4609-8cb0-0ec800e14aad" containerID="976004492a85701ed3f52fa7bc8e5867255ef89dbdafc06b13470404e5d20715" exitCode=0 Oct 11 04:06:28 crc kubenswrapper[4967]: I1011 04:06:28.071525 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"2f52a524-225c-4609-8cb0-0ec800e14aad","Type":"ContainerDied","Data":"976004492a85701ed3f52fa7bc8e5867255ef89dbdafc06b13470404e5d20715"} Oct 11 04:06:28 crc kubenswrapper[4967]: I1011 04:06:28.073501 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6bc7876d45-7pf4q" podUID="b2bf19d4-bf7a-414a-9ad5-f603c3ef5575" containerName="dnsmasq-dns" containerID="cri-o://209853d07705f6072700ccd7046271b186718ad2edbda7dff16d31ad8230b8fa" gracePeriod=10 Oct 11 04:06:28 crc kubenswrapper[4967]: I1011 04:06:28.185450 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-4txc5" podStartSLOduration=3.624821758 podStartE2EDuration="7.185431231s" podCreationTimestamp="2025-10-11 04:06:21 +0000 UTC" firstStartedPulling="2025-10-11 04:06:23.936843704 +0000 UTC m=+911.900052637" lastFinishedPulling="2025-10-11 04:06:27.497453177 +0000 UTC m=+915.460662110" observedRunningTime="2025-10-11 04:06:28.131283773 +0000 UTC m=+916.094492706" watchObservedRunningTime="2025-10-11 04:06:28.185431231 +0000 UTC m=+916.148640174" Oct 11 04:06:28 crc kubenswrapper[4967]: I1011 04:06:28.235018 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-8554648995-56sgg"] Oct 11 04:06:28 crc kubenswrapper[4967]: I1011 04:06:28.537948 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-7pf4q" Oct 11 04:06:28 crc kubenswrapper[4967]: I1011 04:06:28.608978 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2bf19d4-bf7a-414a-9ad5-f603c3ef5575-ovsdbserver-sb\") pod \"b2bf19d4-bf7a-414a-9ad5-f603c3ef5575\" (UID: \"b2bf19d4-bf7a-414a-9ad5-f603c3ef5575\") " Oct 11 04:06:28 crc kubenswrapper[4967]: I1011 04:06:28.609361 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2bf19d4-bf7a-414a-9ad5-f603c3ef5575-dns-svc\") pod \"b2bf19d4-bf7a-414a-9ad5-f603c3ef5575\" (UID: \"b2bf19d4-bf7a-414a-9ad5-f603c3ef5575\") " Oct 11 04:06:28 crc kubenswrapper[4967]: I1011 04:06:28.609489 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-chkqw\" (UniqueName: \"kubernetes.io/projected/b2bf19d4-bf7a-414a-9ad5-f603c3ef5575-kube-api-access-chkqw\") pod \"b2bf19d4-bf7a-414a-9ad5-f603c3ef5575\" (UID: \"b2bf19d4-bf7a-414a-9ad5-f603c3ef5575\") " Oct 11 04:06:28 crc kubenswrapper[4967]: I1011 04:06:28.609578 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2bf19d4-bf7a-414a-9ad5-f603c3ef5575-config\") pod \"b2bf19d4-bf7a-414a-9ad5-f603c3ef5575\" (UID: \"b2bf19d4-bf7a-414a-9ad5-f603c3ef5575\") " Oct 11 04:06:28 crc kubenswrapper[4967]: I1011 04:06:28.615516 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b2bf19d4-bf7a-414a-9ad5-f603c3ef5575-kube-api-access-chkqw" (OuterVolumeSpecName: "kube-api-access-chkqw") pod "b2bf19d4-bf7a-414a-9ad5-f603c3ef5575" (UID: "b2bf19d4-bf7a-414a-9ad5-f603c3ef5575"). InnerVolumeSpecName "kube-api-access-chkqw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:06:28 crc kubenswrapper[4967]: I1011 04:06:28.648094 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2bf19d4-bf7a-414a-9ad5-f603c3ef5575-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b2bf19d4-bf7a-414a-9ad5-f603c3ef5575" (UID: "b2bf19d4-bf7a-414a-9ad5-f603c3ef5575"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:06:28 crc kubenswrapper[4967]: I1011 04:06:28.648887 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2bf19d4-bf7a-414a-9ad5-f603c3ef5575-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "b2bf19d4-bf7a-414a-9ad5-f603c3ef5575" (UID: "b2bf19d4-bf7a-414a-9ad5-f603c3ef5575"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:06:28 crc kubenswrapper[4967]: I1011 04:06:28.651663 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b2bf19d4-bf7a-414a-9ad5-f603c3ef5575-config" (OuterVolumeSpecName: "config") pod "b2bf19d4-bf7a-414a-9ad5-f603c3ef5575" (UID: "b2bf19d4-bf7a-414a-9ad5-f603c3ef5575"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:06:28 crc kubenswrapper[4967]: I1011 04:06:28.693762 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-nb-0" Oct 11 04:06:28 crc kubenswrapper[4967]: I1011 04:06:28.693813 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-nb-0" Oct 11 04:06:28 crc kubenswrapper[4967]: I1011 04:06:28.712017 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b2bf19d4-bf7a-414a-9ad5-f603c3ef5575-config\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:28 crc kubenswrapper[4967]: I1011 04:06:28.712085 4967 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/b2bf19d4-bf7a-414a-9ad5-f603c3ef5575-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:28 crc kubenswrapper[4967]: I1011 04:06:28.712102 4967 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b2bf19d4-bf7a-414a-9ad5-f603c3ef5575-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:28 crc kubenswrapper[4967]: I1011 04:06:28.712115 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-chkqw\" (UniqueName: \"kubernetes.io/projected/b2bf19d4-bf7a-414a-9ad5-f603c3ef5575-kube-api-access-chkqw\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:29 crc kubenswrapper[4967]: I1011 04:06:29.080614 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-galera-0" event={"ID":"ddc18aa7-8599-40c5-a019-972c13162f83","Type":"ContainerStarted","Data":"aa6a17c2e74f137d957e338c43a3317f876f1dc5bf36e581dee420bf678149ae"} Oct 11 04:06:29 crc kubenswrapper[4967]: I1011 04:06:29.089760 4967 generic.go:334] "Generic (PLEG): container finished" podID="bc388f4e-6c72-49ba-a4d4-fb7a01e13681" containerID="78b83ae5a1ce4953c4215fa14f5ff0eebd69b98ba2ca33fe9dd6fe4e0d9c170b" exitCode=0 Oct 11 04:06:29 crc kubenswrapper[4967]: I1011 04:06:29.089862 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-56sgg" event={"ID":"bc388f4e-6c72-49ba-a4d4-fb7a01e13681","Type":"ContainerDied","Data":"78b83ae5a1ce4953c4215fa14f5ff0eebd69b98ba2ca33fe9dd6fe4e0d9c170b"} Oct 11 04:06:29 crc kubenswrapper[4967]: I1011 04:06:29.089893 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-56sgg" event={"ID":"bc388f4e-6c72-49ba-a4d4-fb7a01e13681","Type":"ContainerStarted","Data":"1af19d87a3771c47b86046d8d2da53da2072325f6a8c22cc59616c7344c20658"} Oct 11 04:06:29 crc kubenswrapper[4967]: I1011 04:06:29.095290 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstack-cell1-galera-0" event={"ID":"2f52a524-225c-4609-8cb0-0ec800e14aad","Type":"ContainerStarted","Data":"abc385bcb65b908037454075f5a38c2890178d5dcbf7950c5eae75a18ccf2b6f"} Oct 11 04:06:29 crc kubenswrapper[4967]: I1011 04:06:29.099685 4967 generic.go:334] "Generic (PLEG): container finished" podID="b2bf19d4-bf7a-414a-9ad5-f603c3ef5575" containerID="209853d07705f6072700ccd7046271b186718ad2edbda7dff16d31ad8230b8fa" exitCode=0 Oct 11 04:06:29 crc kubenswrapper[4967]: I1011 04:06:29.100451 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-7pf4q" event={"ID":"b2bf19d4-bf7a-414a-9ad5-f603c3ef5575","Type":"ContainerDied","Data":"209853d07705f6072700ccd7046271b186718ad2edbda7dff16d31ad8230b8fa"} Oct 11 04:06:29 crc kubenswrapper[4967]: I1011 04:06:29.100479 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6bc7876d45-7pf4q" Oct 11 04:06:29 crc kubenswrapper[4967]: I1011 04:06:29.100664 4967 scope.go:117] "RemoveContainer" containerID="209853d07705f6072700ccd7046271b186718ad2edbda7dff16d31ad8230b8fa" Oct 11 04:06:29 crc kubenswrapper[4967]: I1011 04:06:29.100652 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6bc7876d45-7pf4q" event={"ID":"b2bf19d4-bf7a-414a-9ad5-f603c3ef5575","Type":"ContainerDied","Data":"a6d7649366db251a3802e36b053c218fbbc665bce4d593ba52ca6cb332139845"} Oct 11 04:06:29 crc kubenswrapper[4967]: I1011 04:06:29.128719 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-galera-0" podStartSLOduration=18.447050062 podStartE2EDuration="28.128697664s" podCreationTimestamp="2025-10-11 04:06:01 +0000 UTC" firstStartedPulling="2025-10-11 04:06:12.637620567 +0000 UTC m=+900.600829510" lastFinishedPulling="2025-10-11 04:06:22.319268169 +0000 UTC m=+910.282477112" observedRunningTime="2025-10-11 04:06:29.122644501 +0000 UTC m=+917.085853444" watchObservedRunningTime="2025-10-11 04:06:29.128697664 +0000 UTC m=+917.091906597" Oct 11 04:06:29 crc kubenswrapper[4967]: I1011 04:06:29.133107 4967 scope.go:117] "RemoveContainer" containerID="16622120bb6d6340098cb0f262d55c2ced0e00e7f424b93dc1df8f4e5d074fb8" Oct 11 04:06:29 crc kubenswrapper[4967]: I1011 04:06:29.151301 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstack-cell1-galera-0" podStartSLOduration=16.815250593000002 podStartE2EDuration="27.151281517s" podCreationTimestamp="2025-10-11 04:06:02 +0000 UTC" firstStartedPulling="2025-10-11 04:06:12.634834921 +0000 UTC m=+900.598043854" lastFinishedPulling="2025-10-11 04:06:22.970865835 +0000 UTC m=+910.934074778" observedRunningTime="2025-10-11 04:06:29.138358652 +0000 UTC m=+917.101567595" watchObservedRunningTime="2025-10-11 04:06:29.151281517 +0000 UTC m=+917.114490450" Oct 11 04:06:29 crc kubenswrapper[4967]: I1011 04:06:29.161594 4967 scope.go:117] "RemoveContainer" containerID="209853d07705f6072700ccd7046271b186718ad2edbda7dff16d31ad8230b8fa" Oct 11 04:06:29 crc kubenswrapper[4967]: E1011 04:06:29.164559 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"209853d07705f6072700ccd7046271b186718ad2edbda7dff16d31ad8230b8fa\": container with ID starting with 209853d07705f6072700ccd7046271b186718ad2edbda7dff16d31ad8230b8fa not found: ID does not exist" containerID="209853d07705f6072700ccd7046271b186718ad2edbda7dff16d31ad8230b8fa" Oct 11 04:06:29 crc kubenswrapper[4967]: I1011 04:06:29.164606 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"209853d07705f6072700ccd7046271b186718ad2edbda7dff16d31ad8230b8fa"} err="failed to get container status \"209853d07705f6072700ccd7046271b186718ad2edbda7dff16d31ad8230b8fa\": rpc error: code = NotFound desc = could not find container \"209853d07705f6072700ccd7046271b186718ad2edbda7dff16d31ad8230b8fa\": container with ID starting with 209853d07705f6072700ccd7046271b186718ad2edbda7dff16d31ad8230b8fa not found: ID does not exist" Oct 11 04:06:29 crc kubenswrapper[4967]: I1011 04:06:29.164637 4967 scope.go:117] "RemoveContainer" containerID="16622120bb6d6340098cb0f262d55c2ced0e00e7f424b93dc1df8f4e5d074fb8" Oct 11 04:06:29 crc kubenswrapper[4967]: I1011 04:06:29.165560 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ovsdbserver-sb-0" Oct 11 04:06:29 crc kubenswrapper[4967]: E1011 04:06:29.165707 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"16622120bb6d6340098cb0f262d55c2ced0e00e7f424b93dc1df8f4e5d074fb8\": container with ID starting with 16622120bb6d6340098cb0f262d55c2ced0e00e7f424b93dc1df8f4e5d074fb8 not found: ID does not exist" containerID="16622120bb6d6340098cb0f262d55c2ced0e00e7f424b93dc1df8f4e5d074fb8" Oct 11 04:06:29 crc kubenswrapper[4967]: I1011 04:06:29.165751 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"16622120bb6d6340098cb0f262d55c2ced0e00e7f424b93dc1df8f4e5d074fb8"} err="failed to get container status \"16622120bb6d6340098cb0f262d55c2ced0e00e7f424b93dc1df8f4e5d074fb8\": rpc error: code = NotFound desc = could not find container \"16622120bb6d6340098cb0f262d55c2ced0e00e7f424b93dc1df8f4e5d074fb8\": container with ID starting with 16622120bb6d6340098cb0f262d55c2ced0e00e7f424b93dc1df8f4e5d074fb8 not found: ID does not exist" Oct 11 04:06:29 crc kubenswrapper[4967]: I1011 04:06:29.185862 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-7pf4q"] Oct 11 04:06:29 crc kubenswrapper[4967]: I1011 04:06:29.194468 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6bc7876d45-7pf4q"] Oct 11 04:06:29 crc kubenswrapper[4967]: I1011 04:06:29.218916 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-sb-0" Oct 11 04:06:29 crc kubenswrapper[4967]: I1011 04:06:29.619286 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/memcached-0" Oct 11 04:06:30 crc kubenswrapper[4967]: I1011 04:06:30.111057 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-56sgg" event={"ID":"bc388f4e-6c72-49ba-a4d4-fb7a01e13681","Type":"ContainerStarted","Data":"be75bcd98e0eaf7205a5aa56070b3568689c26adaf90b3279e993f5ff3dce78d"} Oct 11 04:06:30 crc kubenswrapper[4967]: I1011 04:06:30.111197 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-8554648995-56sgg" Oct 11 04:06:30 crc kubenswrapper[4967]: I1011 04:06:30.113056 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovsdbserver-sb-0" Oct 11 04:06:30 crc kubenswrapper[4967]: I1011 04:06:30.136509 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-8554648995-56sgg" podStartSLOduration=3.1364919 podStartE2EDuration="3.1364919s" podCreationTimestamp="2025-10-11 04:06:27 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:06:30.132729331 +0000 UTC m=+918.095938264" watchObservedRunningTime="2025-10-11 04:06:30.1364919 +0000 UTC m=+918.099700833" Oct 11 04:06:30 crc kubenswrapper[4967]: I1011 04:06:30.151965 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-sb-0" Oct 11 04:06:30 crc kubenswrapper[4967]: I1011 04:06:30.824672 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b2bf19d4-bf7a-414a-9ad5-f603c3ef5575" path="/var/lib/kubelet/pods/b2bf19d4-bf7a-414a-9ad5-f603c3ef5575/volumes" Oct 11 04:06:31 crc kubenswrapper[4967]: I1011 04:06:31.739237 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ovsdbserver-nb-0" Oct 11 04:06:31 crc kubenswrapper[4967]: I1011 04:06:31.786220 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovsdbserver-nb-0" Oct 11 04:06:32 crc kubenswrapper[4967]: I1011 04:06:32.005916 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-northd-0"] Oct 11 04:06:32 crc kubenswrapper[4967]: E1011 04:06:32.006229 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2bf19d4-bf7a-414a-9ad5-f603c3ef5575" containerName="dnsmasq-dns" Oct 11 04:06:32 crc kubenswrapper[4967]: I1011 04:06:32.006243 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2bf19d4-bf7a-414a-9ad5-f603c3ef5575" containerName="dnsmasq-dns" Oct 11 04:06:32 crc kubenswrapper[4967]: E1011 04:06:32.006256 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b2bf19d4-bf7a-414a-9ad5-f603c3ef5575" containerName="init" Oct 11 04:06:32 crc kubenswrapper[4967]: I1011 04:06:32.006263 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="b2bf19d4-bf7a-414a-9ad5-f603c3ef5575" containerName="init" Oct 11 04:06:32 crc kubenswrapper[4967]: I1011 04:06:32.006424 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="b2bf19d4-bf7a-414a-9ad5-f603c3ef5575" containerName="dnsmasq-dns" Oct 11 04:06:32 crc kubenswrapper[4967]: I1011 04:06:32.007226 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 11 04:06:32 crc kubenswrapper[4967]: W1011 04:06:32.011583 4967 reflector.go:561] object-"openstack"/"cert-ovnnorthd-ovndbs": failed to list *v1.Secret: secrets "cert-ovnnorthd-ovndbs" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 11 04:06:32 crc kubenswrapper[4967]: W1011 04:06:32.011609 4967 reflector.go:561] object-"openstack"/"ovnnorthd-config": failed to list *v1.ConfigMap: configmaps "ovnnorthd-config" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 11 04:06:32 crc kubenswrapper[4967]: E1011 04:06:32.011633 4967 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"cert-ovnnorthd-ovndbs\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"cert-ovnnorthd-ovndbs\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 11 04:06:32 crc kubenswrapper[4967]: E1011 04:06:32.011647 4967 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"ovnnorthd-config\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"ovnnorthd-config\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 11 04:06:32 crc kubenswrapper[4967]: W1011 04:06:32.011680 4967 reflector.go:561] object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-q4pqb": failed to list *v1.Secret: secrets "ovnnorthd-ovnnorthd-dockercfg-q4pqb" is forbidden: User "system:node:crc" cannot list resource "secrets" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 11 04:06:32 crc kubenswrapper[4967]: E1011 04:06:32.011691 4967 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"ovnnorthd-ovnnorthd-dockercfg-q4pqb\": Failed to watch *v1.Secret: failed to list *v1.Secret: secrets \"ovnnorthd-ovnnorthd-dockercfg-q4pqb\" is forbidden: User \"system:node:crc\" cannot list resource \"secrets\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 11 04:06:32 crc kubenswrapper[4967]: W1011 04:06:32.011721 4967 reflector.go:561] object-"openstack"/"ovnnorthd-scripts": failed to list *v1.ConfigMap: configmaps "ovnnorthd-scripts" is forbidden: User "system:node:crc" cannot list resource "configmaps" in API group "" in the namespace "openstack": no relationship found between node 'crc' and this object Oct 11 04:06:32 crc kubenswrapper[4967]: E1011 04:06:32.011733 4967 reflector.go:158] "Unhandled Error" err="object-\"openstack\"/\"ovnnorthd-scripts\": Failed to watch *v1.ConfigMap: failed to list *v1.ConfigMap: configmaps \"ovnnorthd-scripts\" is forbidden: User \"system:node:crc\" cannot list resource \"configmaps\" in API group \"\" in the namespace \"openstack\": no relationship found between node 'crc' and this object" logger="UnhandledError" Oct 11 04:06:32 crc kubenswrapper[4967]: I1011 04:06:32.056882 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 11 04:06:32 crc kubenswrapper[4967]: I1011 04:06:32.085001 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/942f72f2-4ef8-4056-b470-cbc9b2741f10-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"942f72f2-4ef8-4056-b470-cbc9b2741f10\") " pod="openstack/ovn-northd-0" Oct 11 04:06:32 crc kubenswrapper[4967]: I1011 04:06:32.085051 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/942f72f2-4ef8-4056-b470-cbc9b2741f10-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"942f72f2-4ef8-4056-b470-cbc9b2741f10\") " pod="openstack/ovn-northd-0" Oct 11 04:06:32 crc kubenswrapper[4967]: I1011 04:06:32.085121 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjphj\" (UniqueName: \"kubernetes.io/projected/942f72f2-4ef8-4056-b470-cbc9b2741f10-kube-api-access-rjphj\") pod \"ovn-northd-0\" (UID: \"942f72f2-4ef8-4056-b470-cbc9b2741f10\") " pod="openstack/ovn-northd-0" Oct 11 04:06:32 crc kubenswrapper[4967]: I1011 04:06:32.085149 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/942f72f2-4ef8-4056-b470-cbc9b2741f10-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"942f72f2-4ef8-4056-b470-cbc9b2741f10\") " pod="openstack/ovn-northd-0" Oct 11 04:06:32 crc kubenswrapper[4967]: I1011 04:06:32.085200 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/942f72f2-4ef8-4056-b470-cbc9b2741f10-config\") pod \"ovn-northd-0\" (UID: \"942f72f2-4ef8-4056-b470-cbc9b2741f10\") " pod="openstack/ovn-northd-0" Oct 11 04:06:32 crc kubenswrapper[4967]: I1011 04:06:32.085231 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/942f72f2-4ef8-4056-b470-cbc9b2741f10-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"942f72f2-4ef8-4056-b470-cbc9b2741f10\") " pod="openstack/ovn-northd-0" Oct 11 04:06:32 crc kubenswrapper[4967]: I1011 04:06:32.085251 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/942f72f2-4ef8-4056-b470-cbc9b2741f10-scripts\") pod \"ovn-northd-0\" (UID: \"942f72f2-4ef8-4056-b470-cbc9b2741f10\") " pod="openstack/ovn-northd-0" Oct 11 04:06:32 crc kubenswrapper[4967]: I1011 04:06:32.144434 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-4txc5" Oct 11 04:06:32 crc kubenswrapper[4967]: I1011 04:06:32.144498 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-4txc5" Oct 11 04:06:32 crc kubenswrapper[4967]: I1011 04:06:32.187036 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/942f72f2-4ef8-4056-b470-cbc9b2741f10-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"942f72f2-4ef8-4056-b470-cbc9b2741f10\") " pod="openstack/ovn-northd-0" Oct 11 04:06:32 crc kubenswrapper[4967]: I1011 04:06:32.187318 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/942f72f2-4ef8-4056-b470-cbc9b2741f10-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"942f72f2-4ef8-4056-b470-cbc9b2741f10\") " pod="openstack/ovn-northd-0" Oct 11 04:06:32 crc kubenswrapper[4967]: I1011 04:06:32.187357 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjphj\" (UniqueName: \"kubernetes.io/projected/942f72f2-4ef8-4056-b470-cbc9b2741f10-kube-api-access-rjphj\") pod \"ovn-northd-0\" (UID: \"942f72f2-4ef8-4056-b470-cbc9b2741f10\") " pod="openstack/ovn-northd-0" Oct 11 04:06:32 crc kubenswrapper[4967]: I1011 04:06:32.187380 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/942f72f2-4ef8-4056-b470-cbc9b2741f10-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"942f72f2-4ef8-4056-b470-cbc9b2741f10\") " pod="openstack/ovn-northd-0" Oct 11 04:06:32 crc kubenswrapper[4967]: I1011 04:06:32.187417 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/942f72f2-4ef8-4056-b470-cbc9b2741f10-config\") pod \"ovn-northd-0\" (UID: \"942f72f2-4ef8-4056-b470-cbc9b2741f10\") " pod="openstack/ovn-northd-0" Oct 11 04:06:32 crc kubenswrapper[4967]: I1011 04:06:32.187436 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/942f72f2-4ef8-4056-b470-cbc9b2741f10-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"942f72f2-4ef8-4056-b470-cbc9b2741f10\") " pod="openstack/ovn-northd-0" Oct 11 04:06:32 crc kubenswrapper[4967]: I1011 04:06:32.187453 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/942f72f2-4ef8-4056-b470-cbc9b2741f10-scripts\") pod \"ovn-northd-0\" (UID: \"942f72f2-4ef8-4056-b470-cbc9b2741f10\") " pod="openstack/ovn-northd-0" Oct 11 04:06:32 crc kubenswrapper[4967]: I1011 04:06:32.188122 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-rundir\" (UniqueName: \"kubernetes.io/empty-dir/942f72f2-4ef8-4056-b470-cbc9b2741f10-ovn-rundir\") pod \"ovn-northd-0\" (UID: \"942f72f2-4ef8-4056-b470-cbc9b2741f10\") " pod="openstack/ovn-northd-0" Oct 11 04:06:32 crc kubenswrapper[4967]: I1011 04:06:32.194229 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/942f72f2-4ef8-4056-b470-cbc9b2741f10-combined-ca-bundle\") pod \"ovn-northd-0\" (UID: \"942f72f2-4ef8-4056-b470-cbc9b2741f10\") " pod="openstack/ovn-northd-0" Oct 11 04:06:32 crc kubenswrapper[4967]: I1011 04:06:32.195925 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-4txc5" Oct 11 04:06:32 crc kubenswrapper[4967]: I1011 04:06:32.198991 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs-tls-certs\" (UniqueName: \"kubernetes.io/secret/942f72f2-4ef8-4056-b470-cbc9b2741f10-metrics-certs-tls-certs\") pod \"ovn-northd-0\" (UID: \"942f72f2-4ef8-4056-b470-cbc9b2741f10\") " pod="openstack/ovn-northd-0" Oct 11 04:06:32 crc kubenswrapper[4967]: I1011 04:06:32.205540 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjphj\" (UniqueName: \"kubernetes.io/projected/942f72f2-4ef8-4056-b470-cbc9b2741f10-kube-api-access-rjphj\") pod \"ovn-northd-0\" (UID: \"942f72f2-4ef8-4056-b470-cbc9b2741f10\") " pod="openstack/ovn-northd-0" Oct 11 04:06:32 crc kubenswrapper[4967]: I1011 04:06:32.838468 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ovnnorthd-ovndbs" Oct 11 04:06:32 crc kubenswrapper[4967]: I1011 04:06:32.847164 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-northd-tls-certs\" (UniqueName: \"kubernetes.io/secret/942f72f2-4ef8-4056-b470-cbc9b2741f10-ovn-northd-tls-certs\") pod \"ovn-northd-0\" (UID: \"942f72f2-4ef8-4056-b470-cbc9b2741f10\") " pod="openstack/ovn-northd-0" Oct 11 04:06:32 crc kubenswrapper[4967]: I1011 04:06:32.916101 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-galera-0" Oct 11 04:06:32 crc kubenswrapper[4967]: I1011 04:06:32.916427 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-galera-0" Oct 11 04:06:33 crc kubenswrapper[4967]: I1011 04:06:33.004690 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-galera-0" Oct 11 04:06:33 crc kubenswrapper[4967]: I1011 04:06:33.174642 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ovnnorthd-ovnnorthd-dockercfg-q4pqb" Oct 11 04:06:33 crc kubenswrapper[4967]: I1011 04:06:33.175845 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-galera-0" Oct 11 04:06:33 crc kubenswrapper[4967]: I1011 04:06:33.183057 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-scripts" Oct 11 04:06:33 crc kubenswrapper[4967]: I1011 04:06:33.188848 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/942f72f2-4ef8-4056-b470-cbc9b2741f10-scripts\") pod \"ovn-northd-0\" (UID: \"942f72f2-4ef8-4056-b470-cbc9b2741f10\") " pod="openstack/ovn-northd-0" Oct 11 04:06:33 crc kubenswrapper[4967]: E1011 04:06:33.189008 4967 configmap.go:193] Couldn't get configMap openstack/ovnnorthd-config: failed to sync configmap cache: timed out waiting for the condition Oct 11 04:06:33 crc kubenswrapper[4967]: E1011 04:06:33.189234 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/942f72f2-4ef8-4056-b470-cbc9b2741f10-config podName:942f72f2-4ef8-4056-b470-cbc9b2741f10 nodeName:}" failed. No retries permitted until 2025-10-11 04:06:33.689214881 +0000 UTC m=+921.652423824 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "config" (UniqueName: "kubernetes.io/configmap/942f72f2-4ef8-4056-b470-cbc9b2741f10-config") pod "ovn-northd-0" (UID: "942f72f2-4ef8-4056-b470-cbc9b2741f10") : failed to sync configmap cache: timed out waiting for the condition Oct 11 04:06:33 crc kubenswrapper[4967]: I1011 04:06:33.190049 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-4txc5" Oct 11 04:06:33 crc kubenswrapper[4967]: I1011 04:06:33.265720 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4txc5"] Oct 11 04:06:33 crc kubenswrapper[4967]: I1011 04:06:33.467448 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovnnorthd-config" Oct 11 04:06:33 crc kubenswrapper[4967]: I1011 04:06:33.710021 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/942f72f2-4ef8-4056-b470-cbc9b2741f10-config\") pod \"ovn-northd-0\" (UID: \"942f72f2-4ef8-4056-b470-cbc9b2741f10\") " pod="openstack/ovn-northd-0" Oct 11 04:06:33 crc kubenswrapper[4967]: I1011 04:06:33.711320 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/942f72f2-4ef8-4056-b470-cbc9b2741f10-config\") pod \"ovn-northd-0\" (UID: \"942f72f2-4ef8-4056-b470-cbc9b2741f10\") " pod="openstack/ovn-northd-0" Oct 11 04:06:33 crc kubenswrapper[4967]: I1011 04:06:33.845628 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-northd-0" Oct 11 04:06:34 crc kubenswrapper[4967]: I1011 04:06:34.207999 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-northd-0"] Oct 11 04:06:34 crc kubenswrapper[4967]: W1011 04:06:34.212093 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod942f72f2_4ef8_4056_b470_cbc9b2741f10.slice/crio-4b2698fed8016fa56001015f0876529ed408b3c0914ef49704d641c74a2ab2ad WatchSource:0}: Error finding container 4b2698fed8016fa56001015f0876529ed408b3c0914ef49704d641c74a2ab2ad: Status 404 returned error can't find the container with id 4b2698fed8016fa56001015f0876529ed408b3c0914ef49704d641c74a2ab2ad Oct 11 04:06:34 crc kubenswrapper[4967]: I1011 04:06:34.281695 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-create-zpqdx"] Oct 11 04:06:34 crc kubenswrapper[4967]: I1011 04:06:34.282724 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-zpqdx" Oct 11 04:06:34 crc kubenswrapper[4967]: I1011 04:06:34.303864 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:34 crc kubenswrapper[4967]: I1011 04:06:34.303925 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:34 crc kubenswrapper[4967]: I1011 04:06:34.304277 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-zpqdx"] Oct 11 04:06:34 crc kubenswrapper[4967]: I1011 04:06:34.422249 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7q9n4\" (UniqueName: \"kubernetes.io/projected/8f395fae-7337-408a-ae83-1db82f8dc2b3-kube-api-access-7q9n4\") pod \"keystone-db-create-zpqdx\" (UID: \"8f395fae-7337-408a-ae83-1db82f8dc2b3\") " pod="openstack/keystone-db-create-zpqdx" Oct 11 04:06:34 crc kubenswrapper[4967]: I1011 04:06:34.507841 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-create-nfr2c"] Oct 11 04:06:34 crc kubenswrapper[4967]: I1011 04:06:34.509544 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-nfr2c" Oct 11 04:06:34 crc kubenswrapper[4967]: I1011 04:06:34.520691 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-nfr2c"] Oct 11 04:06:34 crc kubenswrapper[4967]: I1011 04:06:34.523273 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7q9n4\" (UniqueName: \"kubernetes.io/projected/8f395fae-7337-408a-ae83-1db82f8dc2b3-kube-api-access-7q9n4\") pod \"keystone-db-create-zpqdx\" (UID: \"8f395fae-7337-408a-ae83-1db82f8dc2b3\") " pod="openstack/keystone-db-create-zpqdx" Oct 11 04:06:34 crc kubenswrapper[4967]: I1011 04:06:34.553966 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7q9n4\" (UniqueName: \"kubernetes.io/projected/8f395fae-7337-408a-ae83-1db82f8dc2b3-kube-api-access-7q9n4\") pod \"keystone-db-create-zpqdx\" (UID: \"8f395fae-7337-408a-ae83-1db82f8dc2b3\") " pod="openstack/keystone-db-create-zpqdx" Oct 11 04:06:34 crc kubenswrapper[4967]: I1011 04:06:34.601116 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-zpqdx" Oct 11 04:06:34 crc kubenswrapper[4967]: I1011 04:06:34.625165 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qsvxq\" (UniqueName: \"kubernetes.io/projected/937176d5-2a41-4b55-89a8-4e2b0019b773-kube-api-access-qsvxq\") pod \"placement-db-create-nfr2c\" (UID: \"937176d5-2a41-4b55-89a8-4e2b0019b773\") " pod="openstack/placement-db-create-nfr2c" Oct 11 04:06:34 crc kubenswrapper[4967]: I1011 04:06:34.727142 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qsvxq\" (UniqueName: \"kubernetes.io/projected/937176d5-2a41-4b55-89a8-4e2b0019b773-kube-api-access-qsvxq\") pod \"placement-db-create-nfr2c\" (UID: \"937176d5-2a41-4b55-89a8-4e2b0019b773\") " pod="openstack/placement-db-create-nfr2c" Oct 11 04:06:34 crc kubenswrapper[4967]: I1011 04:06:34.749936 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qsvxq\" (UniqueName: \"kubernetes.io/projected/937176d5-2a41-4b55-89a8-4e2b0019b773-kube-api-access-qsvxq\") pod \"placement-db-create-nfr2c\" (UID: \"937176d5-2a41-4b55-89a8-4e2b0019b773\") " pod="openstack/placement-db-create-nfr2c" Oct 11 04:06:34 crc kubenswrapper[4967]: I1011 04:06:34.842343 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-nfr2c" Oct 11 04:06:35 crc kubenswrapper[4967]: I1011 04:06:35.016952 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-create-zpqdx"] Oct 11 04:06:35 crc kubenswrapper[4967]: I1011 04:06:35.185139 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-zpqdx" event={"ID":"8f395fae-7337-408a-ae83-1db82f8dc2b3","Type":"ContainerStarted","Data":"e2956ccd87aed6d1fd2ad1dfaf64aa7c1338487c3cae7d5380d8892f8d8621cc"} Oct 11 04:06:35 crc kubenswrapper[4967]: I1011 04:06:35.186694 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"942f72f2-4ef8-4056-b470-cbc9b2741f10","Type":"ContainerStarted","Data":"4b2698fed8016fa56001015f0876529ed408b3c0914ef49704d641c74a2ab2ad"} Oct 11 04:06:35 crc kubenswrapper[4967]: I1011 04:06:35.186989 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-4txc5" podUID="245cf6c7-f572-4395-afdf-ef0795ff6c05" containerName="registry-server" containerID="cri-o://f7b0a3c9e42fbbfdff1690f885565d0619d4e490fd7ccfe45a21c3f0eff56fda" gracePeriod=2 Oct 11 04:06:35 crc kubenswrapper[4967]: I1011 04:06:35.307726 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-create-nfr2c"] Oct 11 04:06:35 crc kubenswrapper[4967]: W1011 04:06:35.377770 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod937176d5_2a41_4b55_89a8_4e2b0019b773.slice/crio-c275ce782d7b6828564f4a41fd6f7f60d5c787b6fad4fdb51d5305b482e0c53b WatchSource:0}: Error finding container c275ce782d7b6828564f4a41fd6f7f60d5c787b6fad4fdb51d5305b482e0c53b: Status 404 returned error can't find the container with id c275ce782d7b6828564f4a41fd6f7f60d5c787b6fad4fdb51d5305b482e0c53b Oct 11 04:06:35 crc kubenswrapper[4967]: E1011 04:06:35.439447 4967 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f395fae_7337_408a_ae83_1db82f8dc2b3.slice/crio-81fd1c2ab306fc3598137e0222bb8f156cbf18904c6b0d8cf76277de06301ba5.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8f395fae_7337_408a_ae83_1db82f8dc2b3.slice/crio-conmon-81fd1c2ab306fc3598137e0222bb8f156cbf18904c6b0d8cf76277de06301ba5.scope\": RecentStats: unable to find data in memory cache]" Oct 11 04:06:35 crc kubenswrapper[4967]: I1011 04:06:35.635064 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4txc5" Oct 11 04:06:35 crc kubenswrapper[4967]: I1011 04:06:35.750262 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-swjq2\" (UniqueName: \"kubernetes.io/projected/245cf6c7-f572-4395-afdf-ef0795ff6c05-kube-api-access-swjq2\") pod \"245cf6c7-f572-4395-afdf-ef0795ff6c05\" (UID: \"245cf6c7-f572-4395-afdf-ef0795ff6c05\") " Oct 11 04:06:35 crc kubenswrapper[4967]: I1011 04:06:35.750425 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/245cf6c7-f572-4395-afdf-ef0795ff6c05-utilities\") pod \"245cf6c7-f572-4395-afdf-ef0795ff6c05\" (UID: \"245cf6c7-f572-4395-afdf-ef0795ff6c05\") " Oct 11 04:06:35 crc kubenswrapper[4967]: I1011 04:06:35.750491 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/245cf6c7-f572-4395-afdf-ef0795ff6c05-catalog-content\") pod \"245cf6c7-f572-4395-afdf-ef0795ff6c05\" (UID: \"245cf6c7-f572-4395-afdf-ef0795ff6c05\") " Oct 11 04:06:35 crc kubenswrapper[4967]: I1011 04:06:35.751470 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/245cf6c7-f572-4395-afdf-ef0795ff6c05-utilities" (OuterVolumeSpecName: "utilities") pod "245cf6c7-f572-4395-afdf-ef0795ff6c05" (UID: "245cf6c7-f572-4395-afdf-ef0795ff6c05"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:06:35 crc kubenswrapper[4967]: I1011 04:06:35.757358 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/245cf6c7-f572-4395-afdf-ef0795ff6c05-kube-api-access-swjq2" (OuterVolumeSpecName: "kube-api-access-swjq2") pod "245cf6c7-f572-4395-afdf-ef0795ff6c05" (UID: "245cf6c7-f572-4395-afdf-ef0795ff6c05"). InnerVolumeSpecName "kube-api-access-swjq2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:06:35 crc kubenswrapper[4967]: I1011 04:06:35.813521 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/245cf6c7-f572-4395-afdf-ef0795ff6c05-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "245cf6c7-f572-4395-afdf-ef0795ff6c05" (UID: "245cf6c7-f572-4395-afdf-ef0795ff6c05"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:06:35 crc kubenswrapper[4967]: I1011 04:06:35.852922 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/245cf6c7-f572-4395-afdf-ef0795ff6c05-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:35 crc kubenswrapper[4967]: I1011 04:06:35.853325 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-swjq2\" (UniqueName: \"kubernetes.io/projected/245cf6c7-f572-4395-afdf-ef0795ff6c05-kube-api-access-swjq2\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:35 crc kubenswrapper[4967]: I1011 04:06:35.853343 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/245cf6c7-f572-4395-afdf-ef0795ff6c05-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.200872 4967 generic.go:334] "Generic (PLEG): container finished" podID="245cf6c7-f572-4395-afdf-ef0795ff6c05" containerID="f7b0a3c9e42fbbfdff1690f885565d0619d4e490fd7ccfe45a21c3f0eff56fda" exitCode=0 Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.200932 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-4txc5" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.200929 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4txc5" event={"ID":"245cf6c7-f572-4395-afdf-ef0795ff6c05","Type":"ContainerDied","Data":"f7b0a3c9e42fbbfdff1690f885565d0619d4e490fd7ccfe45a21c3f0eff56fda"} Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.200996 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-4txc5" event={"ID":"245cf6c7-f572-4395-afdf-ef0795ff6c05","Type":"ContainerDied","Data":"f9f0de980ee68fa7e0fba63fccc20dedba81145da333b2de684c4d5b9246542c"} Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.201018 4967 scope.go:117] "RemoveContainer" containerID="f7b0a3c9e42fbbfdff1690f885565d0619d4e490fd7ccfe45a21c3f0eff56fda" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.203901 4967 generic.go:334] "Generic (PLEG): container finished" podID="937176d5-2a41-4b55-89a8-4e2b0019b773" containerID="a29180c1a7204d0a126ff43042e2ee8b3dd7262ae170a94a670169d29b5a7fd7" exitCode=0 Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.203995 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-nfr2c" event={"ID":"937176d5-2a41-4b55-89a8-4e2b0019b773","Type":"ContainerDied","Data":"a29180c1a7204d0a126ff43042e2ee8b3dd7262ae170a94a670169d29b5a7fd7"} Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.204026 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-nfr2c" event={"ID":"937176d5-2a41-4b55-89a8-4e2b0019b773","Type":"ContainerStarted","Data":"c275ce782d7b6828564f4a41fd6f7f60d5c787b6fad4fdb51d5305b482e0c53b"} Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.207590 4967 generic.go:334] "Generic (PLEG): container finished" podID="8f395fae-7337-408a-ae83-1db82f8dc2b3" containerID="81fd1c2ab306fc3598137e0222bb8f156cbf18904c6b0d8cf76277de06301ba5" exitCode=0 Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.207686 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-zpqdx" event={"ID":"8f395fae-7337-408a-ae83-1db82f8dc2b3","Type":"ContainerDied","Data":"81fd1c2ab306fc3598137e0222bb8f156cbf18904c6b0d8cf76277de06301ba5"} Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.210876 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"942f72f2-4ef8-4056-b470-cbc9b2741f10","Type":"ContainerStarted","Data":"8d089e67e3c6dad2424acf718c06d5b4877dfd42e7be7151f8344f2bda407ac8"} Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.210930 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-northd-0" event={"ID":"942f72f2-4ef8-4056-b470-cbc9b2741f10","Type":"ContainerStarted","Data":"c67a7b2828927094f821624887f1a30cb2e6701e444dd126d424d0e236dff692"} Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.211409 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ovn-northd-0" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.234159 4967 scope.go:117] "RemoveContainer" containerID="d1f8bdb850efc1c71549c5625047537b78b2be4905fbdb7b18043592d1dbfd6f" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.267546 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ovn-northd-0" podStartSLOduration=4.061648443 podStartE2EDuration="5.267519305s" podCreationTimestamp="2025-10-11 04:06:31 +0000 UTC" firstStartedPulling="2025-10-11 04:06:34.213800294 +0000 UTC m=+922.177009227" lastFinishedPulling="2025-10-11 04:06:35.419671156 +0000 UTC m=+923.382880089" observedRunningTime="2025-10-11 04:06:36.263574393 +0000 UTC m=+924.226783326" watchObservedRunningTime="2025-10-11 04:06:36.267519305 +0000 UTC m=+924.230728248" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.272579 4967 scope.go:117] "RemoveContainer" containerID="e3856541bbe583a18caf3cc5114975737812605d997212f79240d0d7c09ca8fa" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.300359 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-4txc5"] Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.307759 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-4txc5"] Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.328627 4967 scope.go:117] "RemoveContainer" containerID="f7b0a3c9e42fbbfdff1690f885565d0619d4e490fd7ccfe45a21c3f0eff56fda" Oct 11 04:06:36 crc kubenswrapper[4967]: E1011 04:06:36.329177 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7b0a3c9e42fbbfdff1690f885565d0619d4e490fd7ccfe45a21c3f0eff56fda\": container with ID starting with f7b0a3c9e42fbbfdff1690f885565d0619d4e490fd7ccfe45a21c3f0eff56fda not found: ID does not exist" containerID="f7b0a3c9e42fbbfdff1690f885565d0619d4e490fd7ccfe45a21c3f0eff56fda" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.329225 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7b0a3c9e42fbbfdff1690f885565d0619d4e490fd7ccfe45a21c3f0eff56fda"} err="failed to get container status \"f7b0a3c9e42fbbfdff1690f885565d0619d4e490fd7ccfe45a21c3f0eff56fda\": rpc error: code = NotFound desc = could not find container \"f7b0a3c9e42fbbfdff1690f885565d0619d4e490fd7ccfe45a21c3f0eff56fda\": container with ID starting with f7b0a3c9e42fbbfdff1690f885565d0619d4e490fd7ccfe45a21c3f0eff56fda not found: ID does not exist" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.329256 4967 scope.go:117] "RemoveContainer" containerID="d1f8bdb850efc1c71549c5625047537b78b2be4905fbdb7b18043592d1dbfd6f" Oct 11 04:06:36 crc kubenswrapper[4967]: E1011 04:06:36.329719 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d1f8bdb850efc1c71549c5625047537b78b2be4905fbdb7b18043592d1dbfd6f\": container with ID starting with d1f8bdb850efc1c71549c5625047537b78b2be4905fbdb7b18043592d1dbfd6f not found: ID does not exist" containerID="d1f8bdb850efc1c71549c5625047537b78b2be4905fbdb7b18043592d1dbfd6f" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.329753 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d1f8bdb850efc1c71549c5625047537b78b2be4905fbdb7b18043592d1dbfd6f"} err="failed to get container status \"d1f8bdb850efc1c71549c5625047537b78b2be4905fbdb7b18043592d1dbfd6f\": rpc error: code = NotFound desc = could not find container \"d1f8bdb850efc1c71549c5625047537b78b2be4905fbdb7b18043592d1dbfd6f\": container with ID starting with d1f8bdb850efc1c71549c5625047537b78b2be4905fbdb7b18043592d1dbfd6f not found: ID does not exist" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.329775 4967 scope.go:117] "RemoveContainer" containerID="e3856541bbe583a18caf3cc5114975737812605d997212f79240d0d7c09ca8fa" Oct 11 04:06:36 crc kubenswrapper[4967]: E1011 04:06:36.330157 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e3856541bbe583a18caf3cc5114975737812605d997212f79240d0d7c09ca8fa\": container with ID starting with e3856541bbe583a18caf3cc5114975737812605d997212f79240d0d7c09ca8fa not found: ID does not exist" containerID="e3856541bbe583a18caf3cc5114975737812605d997212f79240d0d7c09ca8fa" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.330190 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e3856541bbe583a18caf3cc5114975737812605d997212f79240d0d7c09ca8fa"} err="failed to get container status \"e3856541bbe583a18caf3cc5114975737812605d997212f79240d0d7c09ca8fa\": rpc error: code = NotFound desc = could not find container \"e3856541bbe583a18caf3cc5114975737812605d997212f79240d0d7c09ca8fa\": container with ID starting with e3856541bbe583a18caf3cc5114975737812605d997212f79240d0d7c09ca8fa not found: ID does not exist" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.378508 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.421032 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.578360 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-56sgg"] Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.578914 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-8554648995-56sgg" podUID="bc388f4e-6c72-49ba-a4d4-fb7a01e13681" containerName="dnsmasq-dns" containerID="cri-o://be75bcd98e0eaf7205a5aa56070b3568689c26adaf90b3279e993f5ff3dce78d" gracePeriod=10 Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.580273 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-8554648995-56sgg" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.609579 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/openstack-cell1-galera-0" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.622700 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-58pw9"] Oct 11 04:06:36 crc kubenswrapper[4967]: E1011 04:06:36.623018 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="245cf6c7-f572-4395-afdf-ef0795ff6c05" containerName="extract-content" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.623030 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="245cf6c7-f572-4395-afdf-ef0795ff6c05" containerName="extract-content" Oct 11 04:06:36 crc kubenswrapper[4967]: E1011 04:06:36.623044 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="245cf6c7-f572-4395-afdf-ef0795ff6c05" containerName="extract-utilities" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.623050 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="245cf6c7-f572-4395-afdf-ef0795ff6c05" containerName="extract-utilities" Oct 11 04:06:36 crc kubenswrapper[4967]: E1011 04:06:36.623080 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="245cf6c7-f572-4395-afdf-ef0795ff6c05" containerName="registry-server" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.623086 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="245cf6c7-f572-4395-afdf-ef0795ff6c05" containerName="registry-server" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.623242 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="245cf6c7-f572-4395-afdf-ef0795ff6c05" containerName="registry-server" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.623980 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-58pw9" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.677608 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a6cac387-4924-460d-8083-d9f4c296f2a3-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-58pw9\" (UID: \"a6cac387-4924-460d-8083-d9f4c296f2a3\") " pod="openstack/dnsmasq-dns-b8fbc5445-58pw9" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.677661 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6cac387-4924-460d-8083-d9f4c296f2a3-config\") pod \"dnsmasq-dns-b8fbc5445-58pw9\" (UID: \"a6cac387-4924-460d-8083-d9f4c296f2a3\") " pod="openstack/dnsmasq-dns-b8fbc5445-58pw9" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.677691 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a6cac387-4924-460d-8083-d9f4c296f2a3-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-58pw9\" (UID: \"a6cac387-4924-460d-8083-d9f4c296f2a3\") " pod="openstack/dnsmasq-dns-b8fbc5445-58pw9" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.677720 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d2v9n\" (UniqueName: \"kubernetes.io/projected/a6cac387-4924-460d-8083-d9f4c296f2a3-kube-api-access-d2v9n\") pod \"dnsmasq-dns-b8fbc5445-58pw9\" (UID: \"a6cac387-4924-460d-8083-d9f4c296f2a3\") " pod="openstack/dnsmasq-dns-b8fbc5445-58pw9" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.677786 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6cac387-4924-460d-8083-d9f4c296f2a3-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-58pw9\" (UID: \"a6cac387-4924-460d-8083-d9f4c296f2a3\") " pod="openstack/dnsmasq-dns-b8fbc5445-58pw9" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.684164 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-58pw9"] Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.780536 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6cac387-4924-460d-8083-d9f4c296f2a3-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-58pw9\" (UID: \"a6cac387-4924-460d-8083-d9f4c296f2a3\") " pod="openstack/dnsmasq-dns-b8fbc5445-58pw9" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.780616 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a6cac387-4924-460d-8083-d9f4c296f2a3-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-58pw9\" (UID: \"a6cac387-4924-460d-8083-d9f4c296f2a3\") " pod="openstack/dnsmasq-dns-b8fbc5445-58pw9" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.781518 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6cac387-4924-460d-8083-d9f4c296f2a3-config\") pod \"dnsmasq-dns-b8fbc5445-58pw9\" (UID: \"a6cac387-4924-460d-8083-d9f4c296f2a3\") " pod="openstack/dnsmasq-dns-b8fbc5445-58pw9" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.781896 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6cac387-4924-460d-8083-d9f4c296f2a3-config\") pod \"dnsmasq-dns-b8fbc5445-58pw9\" (UID: \"a6cac387-4924-460d-8083-d9f4c296f2a3\") " pod="openstack/dnsmasq-dns-b8fbc5445-58pw9" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.781946 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a6cac387-4924-460d-8083-d9f4c296f2a3-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-58pw9\" (UID: \"a6cac387-4924-460d-8083-d9f4c296f2a3\") " pod="openstack/dnsmasq-dns-b8fbc5445-58pw9" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.781987 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d2v9n\" (UniqueName: \"kubernetes.io/projected/a6cac387-4924-460d-8083-d9f4c296f2a3-kube-api-access-d2v9n\") pod \"dnsmasq-dns-b8fbc5445-58pw9\" (UID: \"a6cac387-4924-460d-8083-d9f4c296f2a3\") " pod="openstack/dnsmasq-dns-b8fbc5445-58pw9" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.782956 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a6cac387-4924-460d-8083-d9f4c296f2a3-ovsdbserver-sb\") pod \"dnsmasq-dns-b8fbc5445-58pw9\" (UID: \"a6cac387-4924-460d-8083-d9f4c296f2a3\") " pod="openstack/dnsmasq-dns-b8fbc5445-58pw9" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.783062 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a6cac387-4924-460d-8083-d9f4c296f2a3-ovsdbserver-nb\") pod \"dnsmasq-dns-b8fbc5445-58pw9\" (UID: \"a6cac387-4924-460d-8083-d9f4c296f2a3\") " pod="openstack/dnsmasq-dns-b8fbc5445-58pw9" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.783734 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6cac387-4924-460d-8083-d9f4c296f2a3-dns-svc\") pod \"dnsmasq-dns-b8fbc5445-58pw9\" (UID: \"a6cac387-4924-460d-8083-d9f4c296f2a3\") " pod="openstack/dnsmasq-dns-b8fbc5445-58pw9" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.806841 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d2v9n\" (UniqueName: \"kubernetes.io/projected/a6cac387-4924-460d-8083-d9f4c296f2a3-kube-api-access-d2v9n\") pod \"dnsmasq-dns-b8fbc5445-58pw9\" (UID: \"a6cac387-4924-460d-8083-d9f4c296f2a3\") " pod="openstack/dnsmasq-dns-b8fbc5445-58pw9" Oct 11 04:06:36 crc kubenswrapper[4967]: I1011 04:06:36.827249 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="245cf6c7-f572-4395-afdf-ef0795ff6c05" path="/var/lib/kubelet/pods/245cf6c7-f572-4395-afdf-ef0795ff6c05/volumes" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.012841 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-58pw9" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.139449 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-56sgg" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.190014 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc388f4e-6c72-49ba-a4d4-fb7a01e13681-ovsdbserver-nb\") pod \"bc388f4e-6c72-49ba-a4d4-fb7a01e13681\" (UID: \"bc388f4e-6c72-49ba-a4d4-fb7a01e13681\") " Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.190138 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc388f4e-6c72-49ba-a4d4-fb7a01e13681-config\") pod \"bc388f4e-6c72-49ba-a4d4-fb7a01e13681\" (UID: \"bc388f4e-6c72-49ba-a4d4-fb7a01e13681\") " Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.190171 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc388f4e-6c72-49ba-a4d4-fb7a01e13681-dns-svc\") pod \"bc388f4e-6c72-49ba-a4d4-fb7a01e13681\" (UID: \"bc388f4e-6c72-49ba-a4d4-fb7a01e13681\") " Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.190213 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svft6\" (UniqueName: \"kubernetes.io/projected/bc388f4e-6c72-49ba-a4d4-fb7a01e13681-kube-api-access-svft6\") pod \"bc388f4e-6c72-49ba-a4d4-fb7a01e13681\" (UID: \"bc388f4e-6c72-49ba-a4d4-fb7a01e13681\") " Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.190362 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc388f4e-6c72-49ba-a4d4-fb7a01e13681-ovsdbserver-sb\") pod \"bc388f4e-6c72-49ba-a4d4-fb7a01e13681\" (UID: \"bc388f4e-6c72-49ba-a4d4-fb7a01e13681\") " Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.204568 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc388f4e-6c72-49ba-a4d4-fb7a01e13681-kube-api-access-svft6" (OuterVolumeSpecName: "kube-api-access-svft6") pod "bc388f4e-6c72-49ba-a4d4-fb7a01e13681" (UID: "bc388f4e-6c72-49ba-a4d4-fb7a01e13681"). InnerVolumeSpecName "kube-api-access-svft6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.239110 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc388f4e-6c72-49ba-a4d4-fb7a01e13681-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "bc388f4e-6c72-49ba-a4d4-fb7a01e13681" (UID: "bc388f4e-6c72-49ba-a4d4-fb7a01e13681"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.241275 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc388f4e-6c72-49ba-a4d4-fb7a01e13681-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "bc388f4e-6c72-49ba-a4d4-fb7a01e13681" (UID: "bc388f4e-6c72-49ba-a4d4-fb7a01e13681"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.242588 4967 generic.go:334] "Generic (PLEG): container finished" podID="bc388f4e-6c72-49ba-a4d4-fb7a01e13681" containerID="be75bcd98e0eaf7205a5aa56070b3568689c26adaf90b3279e993f5ff3dce78d" exitCode=0 Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.242806 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-8554648995-56sgg" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.243900 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-56sgg" event={"ID":"bc388f4e-6c72-49ba-a4d4-fb7a01e13681","Type":"ContainerDied","Data":"be75bcd98e0eaf7205a5aa56070b3568689c26adaf90b3279e993f5ff3dce78d"} Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.243935 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-8554648995-56sgg" event={"ID":"bc388f4e-6c72-49ba-a4d4-fb7a01e13681","Type":"ContainerDied","Data":"1af19d87a3771c47b86046d8d2da53da2072325f6a8c22cc59616c7344c20658"} Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.243958 4967 scope.go:117] "RemoveContainer" containerID="be75bcd98e0eaf7205a5aa56070b3568689c26adaf90b3279e993f5ff3dce78d" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.254218 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc388f4e-6c72-49ba-a4d4-fb7a01e13681-config" (OuterVolumeSpecName: "config") pod "bc388f4e-6c72-49ba-a4d4-fb7a01e13681" (UID: "bc388f4e-6c72-49ba-a4d4-fb7a01e13681"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.293231 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svft6\" (UniqueName: \"kubernetes.io/projected/bc388f4e-6c72-49ba-a4d4-fb7a01e13681-kube-api-access-svft6\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.293262 4967 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/bc388f4e-6c72-49ba-a4d4-fb7a01e13681-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.293271 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/bc388f4e-6c72-49ba-a4d4-fb7a01e13681-config\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.293279 4967 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/bc388f4e-6c72-49ba-a4d4-fb7a01e13681-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.295888 4967 scope.go:117] "RemoveContainer" containerID="78b83ae5a1ce4953c4215fa14f5ff0eebd69b98ba2ca33fe9dd6fe4e0d9c170b" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.306805 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bc388f4e-6c72-49ba-a4d4-fb7a01e13681-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "bc388f4e-6c72-49ba-a4d4-fb7a01e13681" (UID: "bc388f4e-6c72-49ba-a4d4-fb7a01e13681"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.328263 4967 scope.go:117] "RemoveContainer" containerID="be75bcd98e0eaf7205a5aa56070b3568689c26adaf90b3279e993f5ff3dce78d" Oct 11 04:06:37 crc kubenswrapper[4967]: E1011 04:06:37.328774 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"be75bcd98e0eaf7205a5aa56070b3568689c26adaf90b3279e993f5ff3dce78d\": container with ID starting with be75bcd98e0eaf7205a5aa56070b3568689c26adaf90b3279e993f5ff3dce78d not found: ID does not exist" containerID="be75bcd98e0eaf7205a5aa56070b3568689c26adaf90b3279e993f5ff3dce78d" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.328823 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"be75bcd98e0eaf7205a5aa56070b3568689c26adaf90b3279e993f5ff3dce78d"} err="failed to get container status \"be75bcd98e0eaf7205a5aa56070b3568689c26adaf90b3279e993f5ff3dce78d\": rpc error: code = NotFound desc = could not find container \"be75bcd98e0eaf7205a5aa56070b3568689c26adaf90b3279e993f5ff3dce78d\": container with ID starting with be75bcd98e0eaf7205a5aa56070b3568689c26adaf90b3279e993f5ff3dce78d not found: ID does not exist" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.328855 4967 scope.go:117] "RemoveContainer" containerID="78b83ae5a1ce4953c4215fa14f5ff0eebd69b98ba2ca33fe9dd6fe4e0d9c170b" Oct 11 04:06:37 crc kubenswrapper[4967]: E1011 04:06:37.329148 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"78b83ae5a1ce4953c4215fa14f5ff0eebd69b98ba2ca33fe9dd6fe4e0d9c170b\": container with ID starting with 78b83ae5a1ce4953c4215fa14f5ff0eebd69b98ba2ca33fe9dd6fe4e0d9c170b not found: ID does not exist" containerID="78b83ae5a1ce4953c4215fa14f5ff0eebd69b98ba2ca33fe9dd6fe4e0d9c170b" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.329174 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"78b83ae5a1ce4953c4215fa14f5ff0eebd69b98ba2ca33fe9dd6fe4e0d9c170b"} err="failed to get container status \"78b83ae5a1ce4953c4215fa14f5ff0eebd69b98ba2ca33fe9dd6fe4e0d9c170b\": rpc error: code = NotFound desc = could not find container \"78b83ae5a1ce4953c4215fa14f5ff0eebd69b98ba2ca33fe9dd6fe4e0d9c170b\": container with ID starting with 78b83ae5a1ce4953c4215fa14f5ff0eebd69b98ba2ca33fe9dd6fe4e0d9c170b not found: ID does not exist" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.394523 4967 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/bc388f4e-6c72-49ba-a4d4-fb7a01e13681-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.586239 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-8554648995-56sgg"] Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.594218 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-58pw9"] Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.605134 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-8554648995-56sgg"] Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.611376 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-storage-0"] Oct 11 04:06:37 crc kubenswrapper[4967]: E1011 04:06:37.611937 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc388f4e-6c72-49ba-a4d4-fb7a01e13681" containerName="init" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.611971 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc388f4e-6c72-49ba-a4d4-fb7a01e13681" containerName="init" Oct 11 04:06:37 crc kubenswrapper[4967]: E1011 04:06:37.611993 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bc388f4e-6c72-49ba-a4d4-fb7a01e13681" containerName="dnsmasq-dns" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.612000 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="bc388f4e-6c72-49ba-a4d4-fb7a01e13681" containerName="dnsmasq-dns" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.612343 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="bc388f4e-6c72-49ba-a4d4-fb7a01e13681" containerName="dnsmasq-dns" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.620375 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.626742 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-swift-dockercfg-g48rg" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.626967 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-storage-config-data" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.627148 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-conf" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.627280 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-files" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.635133 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.679236 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-zpqdx" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.698124 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7q9n4\" (UniqueName: \"kubernetes.io/projected/8f395fae-7337-408a-ae83-1db82f8dc2b3-kube-api-access-7q9n4\") pod \"8f395fae-7337-408a-ae83-1db82f8dc2b3\" (UID: \"8f395fae-7337-408a-ae83-1db82f8dc2b3\") " Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.698574 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c09e3333-ad4d-495f-8973-7e2d2f0931d4-lock\") pod \"swift-storage-0\" (UID: \"c09e3333-ad4d-495f-8973-7e2d2f0931d4\") " pod="openstack/swift-storage-0" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.698706 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c09e3333-ad4d-495f-8973-7e2d2f0931d4-etc-swift\") pod \"swift-storage-0\" (UID: \"c09e3333-ad4d-495f-8973-7e2d2f0931d4\") " pod="openstack/swift-storage-0" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.698737 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"c09e3333-ad4d-495f-8973-7e2d2f0931d4\") " pod="openstack/swift-storage-0" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.698874 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ljwdg\" (UniqueName: \"kubernetes.io/projected/c09e3333-ad4d-495f-8973-7e2d2f0931d4-kube-api-access-ljwdg\") pod \"swift-storage-0\" (UID: \"c09e3333-ad4d-495f-8973-7e2d2f0931d4\") " pod="openstack/swift-storage-0" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.698936 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c09e3333-ad4d-495f-8973-7e2d2f0931d4-cache\") pod \"swift-storage-0\" (UID: \"c09e3333-ad4d-495f-8973-7e2d2f0931d4\") " pod="openstack/swift-storage-0" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.702881 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f395fae-7337-408a-ae83-1db82f8dc2b3-kube-api-access-7q9n4" (OuterVolumeSpecName: "kube-api-access-7q9n4") pod "8f395fae-7337-408a-ae83-1db82f8dc2b3" (UID: "8f395fae-7337-408a-ae83-1db82f8dc2b3"). InnerVolumeSpecName "kube-api-access-7q9n4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.711109 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-nfr2c" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.799983 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ljwdg\" (UniqueName: \"kubernetes.io/projected/c09e3333-ad4d-495f-8973-7e2d2f0931d4-kube-api-access-ljwdg\") pod \"swift-storage-0\" (UID: \"c09e3333-ad4d-495f-8973-7e2d2f0931d4\") " pod="openstack/swift-storage-0" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.800031 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c09e3333-ad4d-495f-8973-7e2d2f0931d4-cache\") pod \"swift-storage-0\" (UID: \"c09e3333-ad4d-495f-8973-7e2d2f0931d4\") " pod="openstack/swift-storage-0" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.800173 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c09e3333-ad4d-495f-8973-7e2d2f0931d4-lock\") pod \"swift-storage-0\" (UID: \"c09e3333-ad4d-495f-8973-7e2d2f0931d4\") " pod="openstack/swift-storage-0" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.800749 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cache\" (UniqueName: \"kubernetes.io/empty-dir/c09e3333-ad4d-495f-8973-7e2d2f0931d4-cache\") pod \"swift-storage-0\" (UID: \"c09e3333-ad4d-495f-8973-7e2d2f0931d4\") " pod="openstack/swift-storage-0" Oct 11 04:06:37 crc kubenswrapper[4967]: E1011 04:06:37.800930 4967 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 11 04:06:37 crc kubenswrapper[4967]: E1011 04:06:37.800944 4967 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.801143 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"lock\" (UniqueName: \"kubernetes.io/empty-dir/c09e3333-ad4d-495f-8973-7e2d2f0931d4-lock\") pod \"swift-storage-0\" (UID: \"c09e3333-ad4d-495f-8973-7e2d2f0931d4\") " pod="openstack/swift-storage-0" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.801246 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c09e3333-ad4d-495f-8973-7e2d2f0931d4-etc-swift\") pod \"swift-storage-0\" (UID: \"c09e3333-ad4d-495f-8973-7e2d2f0931d4\") " pod="openstack/swift-storage-0" Oct 11 04:06:37 crc kubenswrapper[4967]: E1011 04:06:37.801304 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c09e3333-ad4d-495f-8973-7e2d2f0931d4-etc-swift podName:c09e3333-ad4d-495f-8973-7e2d2f0931d4 nodeName:}" failed. No retries permitted until 2025-10-11 04:06:38.301268771 +0000 UTC m=+926.264477714 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c09e3333-ad4d-495f-8973-7e2d2f0931d4-etc-swift") pod "swift-storage-0" (UID: "c09e3333-ad4d-495f-8973-7e2d2f0931d4") : configmap "swift-ring-files" not found Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.801352 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"c09e3333-ad4d-495f-8973-7e2d2f0931d4\") " pod="openstack/swift-storage-0" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.801482 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7q9n4\" (UniqueName: \"kubernetes.io/projected/8f395fae-7337-408a-ae83-1db82f8dc2b3-kube-api-access-7q9n4\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.801581 4967 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"c09e3333-ad4d-495f-8973-7e2d2f0931d4\") device mount path \"/mnt/openstack/pv09\"" pod="openstack/swift-storage-0" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.821215 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage09-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage09-crc\") pod \"swift-storage-0\" (UID: \"c09e3333-ad4d-495f-8973-7e2d2f0931d4\") " pod="openstack/swift-storage-0" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.822274 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ljwdg\" (UniqueName: \"kubernetes.io/projected/c09e3333-ad4d-495f-8973-7e2d2f0931d4-kube-api-access-ljwdg\") pod \"swift-storage-0\" (UID: \"c09e3333-ad4d-495f-8973-7e2d2f0931d4\") " pod="openstack/swift-storage-0" Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.902825 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qsvxq\" (UniqueName: \"kubernetes.io/projected/937176d5-2a41-4b55-89a8-4e2b0019b773-kube-api-access-qsvxq\") pod \"937176d5-2a41-4b55-89a8-4e2b0019b773\" (UID: \"937176d5-2a41-4b55-89a8-4e2b0019b773\") " Oct 11 04:06:37 crc kubenswrapper[4967]: I1011 04:06:37.905569 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/937176d5-2a41-4b55-89a8-4e2b0019b773-kube-api-access-qsvxq" (OuterVolumeSpecName: "kube-api-access-qsvxq") pod "937176d5-2a41-4b55-89a8-4e2b0019b773" (UID: "937176d5-2a41-4b55-89a8-4e2b0019b773"). InnerVolumeSpecName "kube-api-access-qsvxq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.004632 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qsvxq\" (UniqueName: \"kubernetes.io/projected/937176d5-2a41-4b55-89a8-4e2b0019b773-kube-api-access-qsvxq\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.230138 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-ring-rebalance-6rr4l"] Oct 11 04:06:38 crc kubenswrapper[4967]: E1011 04:06:38.230468 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="937176d5-2a41-4b55-89a8-4e2b0019b773" containerName="mariadb-database-create" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.230482 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="937176d5-2a41-4b55-89a8-4e2b0019b773" containerName="mariadb-database-create" Oct 11 04:06:38 crc kubenswrapper[4967]: E1011 04:06:38.230513 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f395fae-7337-408a-ae83-1db82f8dc2b3" containerName="mariadb-database-create" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.230521 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f395fae-7337-408a-ae83-1db82f8dc2b3" containerName="mariadb-database-create" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.230700 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f395fae-7337-408a-ae83-1db82f8dc2b3" containerName="mariadb-database-create" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.230730 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="937176d5-2a41-4b55-89a8-4e2b0019b773" containerName="mariadb-database-create" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.231352 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6rr4l" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.233131 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.234728 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-scripts" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.237101 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"swift-ring-config-data" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.251833 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-6rr4l"] Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.253839 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-create-zpqdx" event={"ID":"8f395fae-7337-408a-ae83-1db82f8dc2b3","Type":"ContainerDied","Data":"e2956ccd87aed6d1fd2ad1dfaf64aa7c1338487c3cae7d5380d8892f8d8621cc"} Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.253879 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e2956ccd87aed6d1fd2ad1dfaf64aa7c1338487c3cae7d5380d8892f8d8621cc" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.253945 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-create-zpqdx" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.256259 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-create-nfr2c" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.256259 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-create-nfr2c" event={"ID":"937176d5-2a41-4b55-89a8-4e2b0019b773","Type":"ContainerDied","Data":"c275ce782d7b6828564f4a41fd6f7f60d5c787b6fad4fdb51d5305b482e0c53b"} Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.256372 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c275ce782d7b6828564f4a41fd6f7f60d5c787b6fad4fdb51d5305b482e0c53b" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.258463 4967 generic.go:334] "Generic (PLEG): container finished" podID="a6cac387-4924-460d-8083-d9f4c296f2a3" containerID="7eb4833878553650843b57cb3dbfd20b888d8a902999cbe64d1f71b82df556b5" exitCode=0 Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.258676 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-58pw9" event={"ID":"a6cac387-4924-460d-8083-d9f4c296f2a3","Type":"ContainerDied","Data":"7eb4833878553650843b57cb3dbfd20b888d8a902999cbe64d1f71b82df556b5"} Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.258713 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-58pw9" event={"ID":"a6cac387-4924-460d-8083-d9f4c296f2a3","Type":"ContainerStarted","Data":"d4944dab1f7db267d8aeebad3afb684f9385329957f3e65838fcc251386f1c72"} Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.308858 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-ring-data-devices\") pod \"swift-ring-rebalance-6rr4l\" (UID: \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\") " pod="openstack/swift-ring-rebalance-6rr4l" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.309091 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-etc-swift\") pod \"swift-ring-rebalance-6rr4l\" (UID: \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\") " pod="openstack/swift-ring-rebalance-6rr4l" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.309310 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l95j8\" (UniqueName: \"kubernetes.io/projected/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-kube-api-access-l95j8\") pod \"swift-ring-rebalance-6rr4l\" (UID: \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\") " pod="openstack/swift-ring-rebalance-6rr4l" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.309472 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c09e3333-ad4d-495f-8973-7e2d2f0931d4-etc-swift\") pod \"swift-storage-0\" (UID: \"c09e3333-ad4d-495f-8973-7e2d2f0931d4\") " pod="openstack/swift-storage-0" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.310175 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-swiftconf\") pod \"swift-ring-rebalance-6rr4l\" (UID: \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\") " pod="openstack/swift-ring-rebalance-6rr4l" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.310722 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-combined-ca-bundle\") pod \"swift-ring-rebalance-6rr4l\" (UID: \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\") " pod="openstack/swift-ring-rebalance-6rr4l" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.310886 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-scripts\") pod \"swift-ring-rebalance-6rr4l\" (UID: \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\") " pod="openstack/swift-ring-rebalance-6rr4l" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.311055 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-dispersionconf\") pod \"swift-ring-rebalance-6rr4l\" (UID: \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\") " pod="openstack/swift-ring-rebalance-6rr4l" Oct 11 04:06:38 crc kubenswrapper[4967]: E1011 04:06:38.312038 4967 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 11 04:06:38 crc kubenswrapper[4967]: E1011 04:06:38.312171 4967 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 11 04:06:38 crc kubenswrapper[4967]: E1011 04:06:38.312285 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c09e3333-ad4d-495f-8973-7e2d2f0931d4-etc-swift podName:c09e3333-ad4d-495f-8973-7e2d2f0931d4 nodeName:}" failed. No retries permitted until 2025-10-11 04:06:39.312264997 +0000 UTC m=+927.275473930 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c09e3333-ad4d-495f-8973-7e2d2f0931d4-etc-swift") pod "swift-storage-0" (UID: "c09e3333-ad4d-495f-8973-7e2d2f0931d4") : configmap "swift-ring-files" not found Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.412810 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-ring-data-devices\") pod \"swift-ring-rebalance-6rr4l\" (UID: \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\") " pod="openstack/swift-ring-rebalance-6rr4l" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.412855 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-etc-swift\") pod \"swift-ring-rebalance-6rr4l\" (UID: \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\") " pod="openstack/swift-ring-rebalance-6rr4l" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.412878 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l95j8\" (UniqueName: \"kubernetes.io/projected/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-kube-api-access-l95j8\") pod \"swift-ring-rebalance-6rr4l\" (UID: \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\") " pod="openstack/swift-ring-rebalance-6rr4l" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.412932 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-swiftconf\") pod \"swift-ring-rebalance-6rr4l\" (UID: \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\") " pod="openstack/swift-ring-rebalance-6rr4l" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.412965 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-combined-ca-bundle\") pod \"swift-ring-rebalance-6rr4l\" (UID: \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\") " pod="openstack/swift-ring-rebalance-6rr4l" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.412983 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-scripts\") pod \"swift-ring-rebalance-6rr4l\" (UID: \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\") " pod="openstack/swift-ring-rebalance-6rr4l" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.413020 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-dispersionconf\") pod \"swift-ring-rebalance-6rr4l\" (UID: \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\") " pod="openstack/swift-ring-rebalance-6rr4l" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.414434 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-ring-data-devices\") pod \"swift-ring-rebalance-6rr4l\" (UID: \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\") " pod="openstack/swift-ring-rebalance-6rr4l" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.414753 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-etc-swift\") pod \"swift-ring-rebalance-6rr4l\" (UID: \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\") " pod="openstack/swift-ring-rebalance-6rr4l" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.415020 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-scripts\") pod \"swift-ring-rebalance-6rr4l\" (UID: \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\") " pod="openstack/swift-ring-rebalance-6rr4l" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.417462 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-combined-ca-bundle\") pod \"swift-ring-rebalance-6rr4l\" (UID: \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\") " pod="openstack/swift-ring-rebalance-6rr4l" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.417672 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-dispersionconf\") pod \"swift-ring-rebalance-6rr4l\" (UID: \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\") " pod="openstack/swift-ring-rebalance-6rr4l" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.419511 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-swiftconf\") pod \"swift-ring-rebalance-6rr4l\" (UID: \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\") " pod="openstack/swift-ring-rebalance-6rr4l" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.448721 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l95j8\" (UniqueName: \"kubernetes.io/projected/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-kube-api-access-l95j8\") pod \"swift-ring-rebalance-6rr4l\" (UID: \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\") " pod="openstack/swift-ring-rebalance-6rr4l" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.547129 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6rr4l" Oct 11 04:06:38 crc kubenswrapper[4967]: I1011 04:06:38.825127 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc388f4e-6c72-49ba-a4d4-fb7a01e13681" path="/var/lib/kubelet/pods/bc388f4e-6c72-49ba-a4d4-fb7a01e13681/volumes" Oct 11 04:06:39 crc kubenswrapper[4967]: I1011 04:06:39.019371 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-ring-rebalance-6rr4l"] Oct 11 04:06:39 crc kubenswrapper[4967]: W1011 04:06:39.032860 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3f2085d6_0a42_4e1e_b52d_e9379d2a4838.slice/crio-08a13f6849a09fdd5a5a4f60260b83af9b67e0798d1e5393e3a69d13c5b3d8f1 WatchSource:0}: Error finding container 08a13f6849a09fdd5a5a4f60260b83af9b67e0798d1e5393e3a69d13c5b3d8f1: Status 404 returned error can't find the container with id 08a13f6849a09fdd5a5a4f60260b83af9b67e0798d1e5393e3a69d13c5b3d8f1 Oct 11 04:06:39 crc kubenswrapper[4967]: I1011 04:06:39.268909 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-6rr4l" event={"ID":"3f2085d6-0a42-4e1e-b52d-e9379d2a4838","Type":"ContainerStarted","Data":"08a13f6849a09fdd5a5a4f60260b83af9b67e0798d1e5393e3a69d13c5b3d8f1"} Oct 11 04:06:39 crc kubenswrapper[4967]: I1011 04:06:39.271147 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-58pw9" event={"ID":"a6cac387-4924-460d-8083-d9f4c296f2a3","Type":"ContainerStarted","Data":"5f4ca90a58108c90e60b5b8d2945a4fbe89b9bfd3a1e06fb1885314e1f87da87"} Oct 11 04:06:39 crc kubenswrapper[4967]: I1011 04:06:39.271336 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-b8fbc5445-58pw9" Oct 11 04:06:39 crc kubenswrapper[4967]: I1011 04:06:39.309733 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-b8fbc5445-58pw9" podStartSLOduration=3.309713198 podStartE2EDuration="3.309713198s" podCreationTimestamp="2025-10-11 04:06:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:06:39.303950392 +0000 UTC m=+927.267159325" watchObservedRunningTime="2025-10-11 04:06:39.309713198 +0000 UTC m=+927.272922131" Oct 11 04:06:39 crc kubenswrapper[4967]: I1011 04:06:39.326130 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c09e3333-ad4d-495f-8973-7e2d2f0931d4-etc-swift\") pod \"swift-storage-0\" (UID: \"c09e3333-ad4d-495f-8973-7e2d2f0931d4\") " pod="openstack/swift-storage-0" Oct 11 04:06:39 crc kubenswrapper[4967]: E1011 04:06:39.326366 4967 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 11 04:06:39 crc kubenswrapper[4967]: E1011 04:06:39.326403 4967 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 11 04:06:39 crc kubenswrapper[4967]: E1011 04:06:39.326455 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c09e3333-ad4d-495f-8973-7e2d2f0931d4-etc-swift podName:c09e3333-ad4d-495f-8973-7e2d2f0931d4 nodeName:}" failed. No retries permitted until 2025-10-11 04:06:41.326436873 +0000 UTC m=+929.289645806 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c09e3333-ad4d-495f-8973-7e2d2f0931d4-etc-swift") pod "swift-storage-0" (UID: "c09e3333-ad4d-495f-8973-7e2d2f0931d4") : configmap "swift-ring-files" not found Oct 11 04:06:39 crc kubenswrapper[4967]: I1011 04:06:39.820183 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-create-jrjv4"] Oct 11 04:06:39 crc kubenswrapper[4967]: I1011 04:06:39.822534 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-jrjv4" Oct 11 04:06:39 crc kubenswrapper[4967]: I1011 04:06:39.832962 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-jrjv4"] Oct 11 04:06:39 crc kubenswrapper[4967]: I1011 04:06:39.835420 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqhrd\" (UniqueName: \"kubernetes.io/projected/b21fd878-e30e-498f-88f3-8a4f278ee0ec-kube-api-access-dqhrd\") pod \"glance-db-create-jrjv4\" (UID: \"b21fd878-e30e-498f-88f3-8a4f278ee0ec\") " pod="openstack/glance-db-create-jrjv4" Oct 11 04:06:39 crc kubenswrapper[4967]: I1011 04:06:39.937361 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqhrd\" (UniqueName: \"kubernetes.io/projected/b21fd878-e30e-498f-88f3-8a4f278ee0ec-kube-api-access-dqhrd\") pod \"glance-db-create-jrjv4\" (UID: \"b21fd878-e30e-498f-88f3-8a4f278ee0ec\") " pod="openstack/glance-db-create-jrjv4" Oct 11 04:06:39 crc kubenswrapper[4967]: I1011 04:06:39.964021 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqhrd\" (UniqueName: \"kubernetes.io/projected/b21fd878-e30e-498f-88f3-8a4f278ee0ec-kube-api-access-dqhrd\") pod \"glance-db-create-jrjv4\" (UID: \"b21fd878-e30e-498f-88f3-8a4f278ee0ec\") " pod="openstack/glance-db-create-jrjv4" Oct 11 04:06:40 crc kubenswrapper[4967]: I1011 04:06:40.141003 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-jrjv4" Oct 11 04:06:40 crc kubenswrapper[4967]: I1011 04:06:40.584995 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-create-jrjv4"] Oct 11 04:06:41 crc kubenswrapper[4967]: I1011 04:06:41.368092 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c09e3333-ad4d-495f-8973-7e2d2f0931d4-etc-swift\") pod \"swift-storage-0\" (UID: \"c09e3333-ad4d-495f-8973-7e2d2f0931d4\") " pod="openstack/swift-storage-0" Oct 11 04:06:41 crc kubenswrapper[4967]: E1011 04:06:41.368364 4967 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 11 04:06:41 crc kubenswrapper[4967]: E1011 04:06:41.368377 4967 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 11 04:06:41 crc kubenswrapper[4967]: E1011 04:06:41.368428 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c09e3333-ad4d-495f-8973-7e2d2f0931d4-etc-swift podName:c09e3333-ad4d-495f-8973-7e2d2f0931d4 nodeName:}" failed. No retries permitted until 2025-10-11 04:06:45.368412249 +0000 UTC m=+933.331621192 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c09e3333-ad4d-495f-8973-7e2d2f0931d4-etc-swift") pod "swift-storage-0" (UID: "c09e3333-ad4d-495f-8973-7e2d2f0931d4") : configmap "swift-ring-files" not found Oct 11 04:06:44 crc kubenswrapper[4967]: I1011 04:06:44.089562 4967 patch_prober.go:28] interesting pod/machine-config-daemon-sgrv2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:06:44 crc kubenswrapper[4967]: I1011 04:06:44.089962 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:06:44 crc kubenswrapper[4967]: I1011 04:06:44.090019 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" Oct 11 04:06:44 crc kubenswrapper[4967]: I1011 04:06:44.090768 4967 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"f1443b341ff7c816afffce3e40de74a6868efcc5c30c2eb7be83fc5ffc5860c7"} pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 11 04:06:44 crc kubenswrapper[4967]: I1011 04:06:44.090837 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" containerID="cri-o://f1443b341ff7c816afffce3e40de74a6868efcc5c30c2eb7be83fc5ffc5860c7" gracePeriod=600 Oct 11 04:06:44 crc kubenswrapper[4967]: I1011 04:06:44.279860 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-87d8-account-create-9sfk2"] Oct 11 04:06:44 crc kubenswrapper[4967]: I1011 04:06:44.281177 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-87d8-account-create-9sfk2" Oct 11 04:06:44 crc kubenswrapper[4967]: I1011 04:06:44.284402 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-db-secret" Oct 11 04:06:44 crc kubenswrapper[4967]: I1011 04:06:44.297470 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-87d8-account-create-9sfk2"] Oct 11 04:06:44 crc kubenswrapper[4967]: I1011 04:06:44.421692 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-57sts\" (UniqueName: \"kubernetes.io/projected/35ca69d7-b265-4ab4-a047-5fbdb349f13d-kube-api-access-57sts\") pod \"keystone-87d8-account-create-9sfk2\" (UID: \"35ca69d7-b265-4ab4-a047-5fbdb349f13d\") " pod="openstack/keystone-87d8-account-create-9sfk2" Oct 11 04:06:44 crc kubenswrapper[4967]: I1011 04:06:44.480214 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-d5f2-account-create-dfps4"] Oct 11 04:06:44 crc kubenswrapper[4967]: I1011 04:06:44.481565 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d5f2-account-create-dfps4" Oct 11 04:06:44 crc kubenswrapper[4967]: I1011 04:06:44.484310 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-db-secret" Oct 11 04:06:44 crc kubenswrapper[4967]: I1011 04:06:44.487285 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-d5f2-account-create-dfps4"] Oct 11 04:06:44 crc kubenswrapper[4967]: I1011 04:06:44.523219 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qcxk7\" (UniqueName: \"kubernetes.io/projected/292067b2-43e9-46a1-8940-08bca18ce4f0-kube-api-access-qcxk7\") pod \"placement-d5f2-account-create-dfps4\" (UID: \"292067b2-43e9-46a1-8940-08bca18ce4f0\") " pod="openstack/placement-d5f2-account-create-dfps4" Oct 11 04:06:44 crc kubenswrapper[4967]: I1011 04:06:44.523316 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-57sts\" (UniqueName: \"kubernetes.io/projected/35ca69d7-b265-4ab4-a047-5fbdb349f13d-kube-api-access-57sts\") pod \"keystone-87d8-account-create-9sfk2\" (UID: \"35ca69d7-b265-4ab4-a047-5fbdb349f13d\") " pod="openstack/keystone-87d8-account-create-9sfk2" Oct 11 04:06:44 crc kubenswrapper[4967]: I1011 04:06:44.542311 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-57sts\" (UniqueName: \"kubernetes.io/projected/35ca69d7-b265-4ab4-a047-5fbdb349f13d-kube-api-access-57sts\") pod \"keystone-87d8-account-create-9sfk2\" (UID: \"35ca69d7-b265-4ab4-a047-5fbdb349f13d\") " pod="openstack/keystone-87d8-account-create-9sfk2" Oct 11 04:06:44 crc kubenswrapper[4967]: I1011 04:06:44.601904 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-87d8-account-create-9sfk2" Oct 11 04:06:44 crc kubenswrapper[4967]: I1011 04:06:44.624374 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qcxk7\" (UniqueName: \"kubernetes.io/projected/292067b2-43e9-46a1-8940-08bca18ce4f0-kube-api-access-qcxk7\") pod \"placement-d5f2-account-create-dfps4\" (UID: \"292067b2-43e9-46a1-8940-08bca18ce4f0\") " pod="openstack/placement-d5f2-account-create-dfps4" Oct 11 04:06:44 crc kubenswrapper[4967]: I1011 04:06:44.646325 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qcxk7\" (UniqueName: \"kubernetes.io/projected/292067b2-43e9-46a1-8940-08bca18ce4f0-kube-api-access-qcxk7\") pod \"placement-d5f2-account-create-dfps4\" (UID: \"292067b2-43e9-46a1-8940-08bca18ce4f0\") " pod="openstack/placement-d5f2-account-create-dfps4" Oct 11 04:06:44 crc kubenswrapper[4967]: I1011 04:06:44.801340 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d5f2-account-create-dfps4" Oct 11 04:06:46 crc kubenswrapper[4967]: I1011 04:06:45.439537 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c09e3333-ad4d-495f-8973-7e2d2f0931d4-etc-swift\") pod \"swift-storage-0\" (UID: \"c09e3333-ad4d-495f-8973-7e2d2f0931d4\") " pod="openstack/swift-storage-0" Oct 11 04:06:46 crc kubenswrapper[4967]: E1011 04:06:45.439736 4967 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 11 04:06:46 crc kubenswrapper[4967]: E1011 04:06:45.440057 4967 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 11 04:06:46 crc kubenswrapper[4967]: E1011 04:06:45.440196 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c09e3333-ad4d-495f-8973-7e2d2f0931d4-etc-swift podName:c09e3333-ad4d-495f-8973-7e2d2f0931d4 nodeName:}" failed. No retries permitted until 2025-10-11 04:06:53.440165691 +0000 UTC m=+941.403374634 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c09e3333-ad4d-495f-8973-7e2d2f0931d4-etc-swift") pod "swift-storage-0" (UID: "c09e3333-ad4d-495f-8973-7e2d2f0931d4") : configmap "swift-ring-files" not found Oct 11 04:06:46 crc kubenswrapper[4967]: W1011 04:06:46.813882 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podb21fd878_e30e_498f_88f3_8a4f278ee0ec.slice/crio-26b6cc7ecc0654b65421f95287eafebaa1864d53012043242f340eb159c8cccd WatchSource:0}: Error finding container 26b6cc7ecc0654b65421f95287eafebaa1864d53012043242f340eb159c8cccd: Status 404 returned error can't find the container with id 26b6cc7ecc0654b65421f95287eafebaa1864d53012043242f340eb159c8cccd Oct 11 04:06:47 crc kubenswrapper[4967]: I1011 04:06:47.015263 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-b8fbc5445-58pw9" Oct 11 04:06:47 crc kubenswrapper[4967]: I1011 04:06:47.103214 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-vc6g6"] Oct 11 04:06:47 crc kubenswrapper[4967]: I1011 04:06:47.104605 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-5ccc8479f9-vc6g6" podUID="b9b792fe-ab24-4587-a1f6-2ad194c9eb83" containerName="dnsmasq-dns" containerID="cri-o://da23146880d15c1cb30395f5700e1386827eb8f1150aa9c496aa9c35382d0776" gracePeriod=10 Oct 11 04:06:47 crc kubenswrapper[4967]: I1011 04:06:47.355168 4967 generic.go:334] "Generic (PLEG): container finished" podID="97c07678-14be-410c-b61f-498cb49bc960" containerID="f1443b341ff7c816afffce3e40de74a6868efcc5c30c2eb7be83fc5ffc5860c7" exitCode=0 Oct 11 04:06:47 crc kubenswrapper[4967]: I1011 04:06:47.355236 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" event={"ID":"97c07678-14be-410c-b61f-498cb49bc960","Type":"ContainerDied","Data":"f1443b341ff7c816afffce3e40de74a6868efcc5c30c2eb7be83fc5ffc5860c7"} Oct 11 04:06:47 crc kubenswrapper[4967]: I1011 04:06:47.355347 4967 scope.go:117] "RemoveContainer" containerID="fab92459036250055aee13eec5e67ffc42b24499db86755378bf85f3ee4b07d2" Oct 11 04:06:47 crc kubenswrapper[4967]: I1011 04:06:47.358012 4967 generic.go:334] "Generic (PLEG): container finished" podID="b9b792fe-ab24-4587-a1f6-2ad194c9eb83" containerID="da23146880d15c1cb30395f5700e1386827eb8f1150aa9c496aa9c35382d0776" exitCode=0 Oct 11 04:06:47 crc kubenswrapper[4967]: I1011 04:06:47.358086 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-vc6g6" event={"ID":"b9b792fe-ab24-4587-a1f6-2ad194c9eb83","Type":"ContainerDied","Data":"da23146880d15c1cb30395f5700e1386827eb8f1150aa9c496aa9c35382d0776"} Oct 11 04:06:47 crc kubenswrapper[4967]: I1011 04:06:47.360525 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-jrjv4" event={"ID":"b21fd878-e30e-498f-88f3-8a4f278ee0ec","Type":"ContainerStarted","Data":"26b6cc7ecc0654b65421f95287eafebaa1864d53012043242f340eb159c8cccd"} Oct 11 04:06:48 crc kubenswrapper[4967]: I1011 04:06:48.533187 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-vc6g6" Oct 11 04:06:48 crc kubenswrapper[4967]: I1011 04:06:48.609378 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9b792fe-ab24-4587-a1f6-2ad194c9eb83-dns-svc\") pod \"b9b792fe-ab24-4587-a1f6-2ad194c9eb83\" (UID: \"b9b792fe-ab24-4587-a1f6-2ad194c9eb83\") " Oct 11 04:06:48 crc kubenswrapper[4967]: I1011 04:06:48.609500 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9b792fe-ab24-4587-a1f6-2ad194c9eb83-config\") pod \"b9b792fe-ab24-4587-a1f6-2ad194c9eb83\" (UID: \"b9b792fe-ab24-4587-a1f6-2ad194c9eb83\") " Oct 11 04:06:48 crc kubenswrapper[4967]: I1011 04:06:48.609555 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jdbwh\" (UniqueName: \"kubernetes.io/projected/b9b792fe-ab24-4587-a1f6-2ad194c9eb83-kube-api-access-jdbwh\") pod \"b9b792fe-ab24-4587-a1f6-2ad194c9eb83\" (UID: \"b9b792fe-ab24-4587-a1f6-2ad194c9eb83\") " Oct 11 04:06:48 crc kubenswrapper[4967]: I1011 04:06:48.614463 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9b792fe-ab24-4587-a1f6-2ad194c9eb83-kube-api-access-jdbwh" (OuterVolumeSpecName: "kube-api-access-jdbwh") pod "b9b792fe-ab24-4587-a1f6-2ad194c9eb83" (UID: "b9b792fe-ab24-4587-a1f6-2ad194c9eb83"). InnerVolumeSpecName "kube-api-access-jdbwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:06:48 crc kubenswrapper[4967]: I1011 04:06:48.646324 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9b792fe-ab24-4587-a1f6-2ad194c9eb83-config" (OuterVolumeSpecName: "config") pod "b9b792fe-ab24-4587-a1f6-2ad194c9eb83" (UID: "b9b792fe-ab24-4587-a1f6-2ad194c9eb83"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:06:48 crc kubenswrapper[4967]: I1011 04:06:48.650344 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b9b792fe-ab24-4587-a1f6-2ad194c9eb83-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "b9b792fe-ab24-4587-a1f6-2ad194c9eb83" (UID: "b9b792fe-ab24-4587-a1f6-2ad194c9eb83"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:06:48 crc kubenswrapper[4967]: I1011 04:06:48.663394 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-d5f2-account-create-dfps4"] Oct 11 04:06:48 crc kubenswrapper[4967]: W1011 04:06:48.666060 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod292067b2_43e9_46a1_8940_08bca18ce4f0.slice/crio-3301fecb776fcf18cf5bfa739aa5eb47335f435f607270d6dcc8135d38e96af6 WatchSource:0}: Error finding container 3301fecb776fcf18cf5bfa739aa5eb47335f435f607270d6dcc8135d38e96af6: Status 404 returned error can't find the container with id 3301fecb776fcf18cf5bfa739aa5eb47335f435f607270d6dcc8135d38e96af6 Oct 11 04:06:48 crc kubenswrapper[4967]: I1011 04:06:48.711793 4967 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/b9b792fe-ab24-4587-a1f6-2ad194c9eb83-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:48 crc kubenswrapper[4967]: I1011 04:06:48.712051 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b9b792fe-ab24-4587-a1f6-2ad194c9eb83-config\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:48 crc kubenswrapper[4967]: I1011 04:06:48.712062 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jdbwh\" (UniqueName: \"kubernetes.io/projected/b9b792fe-ab24-4587-a1f6-2ad194c9eb83-kube-api-access-jdbwh\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:48 crc kubenswrapper[4967]: I1011 04:06:48.749341 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-87d8-account-create-9sfk2"] Oct 11 04:06:48 crc kubenswrapper[4967]: W1011 04:06:48.760389 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod35ca69d7_b265_4ab4_a047_5fbdb349f13d.slice/crio-7744a2457109574e0129bf874b0b6a3c1357f33b3678f5859499294749a70a0e WatchSource:0}: Error finding container 7744a2457109574e0129bf874b0b6a3c1357f33b3678f5859499294749a70a0e: Status 404 returned error can't find the container with id 7744a2457109574e0129bf874b0b6a3c1357f33b3678f5859499294749a70a0e Oct 11 04:06:48 crc kubenswrapper[4967]: I1011 04:06:48.907802 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-northd-0" Oct 11 04:06:49 crc kubenswrapper[4967]: I1011 04:06:49.394911 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-6rr4l" event={"ID":"3f2085d6-0a42-4e1e-b52d-e9379d2a4838","Type":"ContainerStarted","Data":"0b09663baaa990bafaf876c5dd5d886e11ab30a188f37d3af2f6327d91da9462"} Oct 11 04:06:49 crc kubenswrapper[4967]: I1011 04:06:49.396630 4967 generic.go:334] "Generic (PLEG): container finished" podID="35ca69d7-b265-4ab4-a047-5fbdb349f13d" containerID="1ae031598cabc6c10cadd267df2488dad0c87a931f779e39bab402b4c50ce80b" exitCode=0 Oct 11 04:06:49 crc kubenswrapper[4967]: I1011 04:06:49.396706 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-87d8-account-create-9sfk2" event={"ID":"35ca69d7-b265-4ab4-a047-5fbdb349f13d","Type":"ContainerDied","Data":"1ae031598cabc6c10cadd267df2488dad0c87a931f779e39bab402b4c50ce80b"} Oct 11 04:06:49 crc kubenswrapper[4967]: I1011 04:06:49.396752 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-87d8-account-create-9sfk2" event={"ID":"35ca69d7-b265-4ab4-a047-5fbdb349f13d","Type":"ContainerStarted","Data":"7744a2457109574e0129bf874b0b6a3c1357f33b3678f5859499294749a70a0e"} Oct 11 04:06:49 crc kubenswrapper[4967]: I1011 04:06:49.398752 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-5ccc8479f9-vc6g6" event={"ID":"b9b792fe-ab24-4587-a1f6-2ad194c9eb83","Type":"ContainerDied","Data":"8e230d575816b5e3aa3bffe002ae98f8e2d23ac163f6e95713ce6d069fe9be0f"} Oct 11 04:06:49 crc kubenswrapper[4967]: I1011 04:06:49.398764 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-5ccc8479f9-vc6g6" Oct 11 04:06:49 crc kubenswrapper[4967]: I1011 04:06:49.398812 4967 scope.go:117] "RemoveContainer" containerID="da23146880d15c1cb30395f5700e1386827eb8f1150aa9c496aa9c35382d0776" Oct 11 04:06:49 crc kubenswrapper[4967]: I1011 04:06:49.402635 4967 generic.go:334] "Generic (PLEG): container finished" podID="b21fd878-e30e-498f-88f3-8a4f278ee0ec" containerID="5db53328505e2416c3f4d98b1042bc99fe763631c2946f9777368c2f0241df8c" exitCode=0 Oct 11 04:06:49 crc kubenswrapper[4967]: I1011 04:06:49.402783 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-jrjv4" event={"ID":"b21fd878-e30e-498f-88f3-8a4f278ee0ec","Type":"ContainerDied","Data":"5db53328505e2416c3f4d98b1042bc99fe763631c2946f9777368c2f0241df8c"} Oct 11 04:06:49 crc kubenswrapper[4967]: I1011 04:06:49.417421 4967 generic.go:334] "Generic (PLEG): container finished" podID="292067b2-43e9-46a1-8940-08bca18ce4f0" containerID="2a41c7134fd43de5cbd967564c77fb546d537bef1bb8b9b99d570a3b67c21e9c" exitCode=0 Oct 11 04:06:49 crc kubenswrapper[4967]: I1011 04:06:49.417504 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d5f2-account-create-dfps4" event={"ID":"292067b2-43e9-46a1-8940-08bca18ce4f0","Type":"ContainerDied","Data":"2a41c7134fd43de5cbd967564c77fb546d537bef1bb8b9b99d570a3b67c21e9c"} Oct 11 04:06:49 crc kubenswrapper[4967]: I1011 04:06:49.417532 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d5f2-account-create-dfps4" event={"ID":"292067b2-43e9-46a1-8940-08bca18ce4f0","Type":"ContainerStarted","Data":"3301fecb776fcf18cf5bfa739aa5eb47335f435f607270d6dcc8135d38e96af6"} Oct 11 04:06:49 crc kubenswrapper[4967]: I1011 04:06:49.445287 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" event={"ID":"97c07678-14be-410c-b61f-498cb49bc960","Type":"ContainerStarted","Data":"37a6f451ae95b67f0b65e3cc023ede1c1770c6a148d14dd88cd2fe303a4c8469"} Oct 11 04:06:49 crc kubenswrapper[4967]: I1011 04:06:49.447137 4967 scope.go:117] "RemoveContainer" containerID="e84fbc857ef08fe10ca354b61edee337ef79228822f2db1fad01f24f50ee52b0" Oct 11 04:06:49 crc kubenswrapper[4967]: I1011 04:06:49.483018 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-ring-rebalance-6rr4l" podStartSLOduration=2.313893609 podStartE2EDuration="11.48299376s" podCreationTimestamp="2025-10-11 04:06:38 +0000 UTC" firstStartedPulling="2025-10-11 04:06:39.03615838 +0000 UTC m=+926.999367313" lastFinishedPulling="2025-10-11 04:06:48.205258521 +0000 UTC m=+936.168467464" observedRunningTime="2025-10-11 04:06:49.424647072 +0000 UTC m=+937.387856005" watchObservedRunningTime="2025-10-11 04:06:49.48299376 +0000 UTC m=+937.446202693" Oct 11 04:06:49 crc kubenswrapper[4967]: I1011 04:06:49.523499 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-vc6g6"] Oct 11 04:06:49 crc kubenswrapper[4967]: I1011 04:06:49.536193 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-5ccc8479f9-vc6g6"] Oct 11 04:06:50 crc kubenswrapper[4967]: I1011 04:06:50.824133 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9b792fe-ab24-4587-a1f6-2ad194c9eb83" path="/var/lib/kubelet/pods/b9b792fe-ab24-4587-a1f6-2ad194c9eb83/volumes" Oct 11 04:06:50 crc kubenswrapper[4967]: I1011 04:06:50.825174 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-87d8-account-create-9sfk2" Oct 11 04:06:50 crc kubenswrapper[4967]: I1011 04:06:50.848662 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-57sts\" (UniqueName: \"kubernetes.io/projected/35ca69d7-b265-4ab4-a047-5fbdb349f13d-kube-api-access-57sts\") pod \"35ca69d7-b265-4ab4-a047-5fbdb349f13d\" (UID: \"35ca69d7-b265-4ab4-a047-5fbdb349f13d\") " Oct 11 04:06:50 crc kubenswrapper[4967]: I1011 04:06:50.855688 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35ca69d7-b265-4ab4-a047-5fbdb349f13d-kube-api-access-57sts" (OuterVolumeSpecName: "kube-api-access-57sts") pod "35ca69d7-b265-4ab4-a047-5fbdb349f13d" (UID: "35ca69d7-b265-4ab4-a047-5fbdb349f13d"). InnerVolumeSpecName "kube-api-access-57sts". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:06:50 crc kubenswrapper[4967]: I1011 04:06:50.930627 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d5f2-account-create-dfps4" Oct 11 04:06:50 crc kubenswrapper[4967]: I1011 04:06:50.937216 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-jrjv4" Oct 11 04:06:50 crc kubenswrapper[4967]: I1011 04:06:50.950845 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqhrd\" (UniqueName: \"kubernetes.io/projected/b21fd878-e30e-498f-88f3-8a4f278ee0ec-kube-api-access-dqhrd\") pod \"b21fd878-e30e-498f-88f3-8a4f278ee0ec\" (UID: \"b21fd878-e30e-498f-88f3-8a4f278ee0ec\") " Oct 11 04:06:50 crc kubenswrapper[4967]: I1011 04:06:50.950977 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qcxk7\" (UniqueName: \"kubernetes.io/projected/292067b2-43e9-46a1-8940-08bca18ce4f0-kube-api-access-qcxk7\") pod \"292067b2-43e9-46a1-8940-08bca18ce4f0\" (UID: \"292067b2-43e9-46a1-8940-08bca18ce4f0\") " Oct 11 04:06:50 crc kubenswrapper[4967]: I1011 04:06:50.951371 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-57sts\" (UniqueName: \"kubernetes.io/projected/35ca69d7-b265-4ab4-a047-5fbdb349f13d-kube-api-access-57sts\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:50 crc kubenswrapper[4967]: I1011 04:06:50.954291 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b21fd878-e30e-498f-88f3-8a4f278ee0ec-kube-api-access-dqhrd" (OuterVolumeSpecName: "kube-api-access-dqhrd") pod "b21fd878-e30e-498f-88f3-8a4f278ee0ec" (UID: "b21fd878-e30e-498f-88f3-8a4f278ee0ec"). InnerVolumeSpecName "kube-api-access-dqhrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:06:50 crc kubenswrapper[4967]: I1011 04:06:50.957581 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/292067b2-43e9-46a1-8940-08bca18ce4f0-kube-api-access-qcxk7" (OuterVolumeSpecName: "kube-api-access-qcxk7") pod "292067b2-43e9-46a1-8940-08bca18ce4f0" (UID: "292067b2-43e9-46a1-8940-08bca18ce4f0"). InnerVolumeSpecName "kube-api-access-qcxk7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:06:51 crc kubenswrapper[4967]: I1011 04:06:51.052793 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqhrd\" (UniqueName: \"kubernetes.io/projected/b21fd878-e30e-498f-88f3-8a4f278ee0ec-kube-api-access-dqhrd\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:51 crc kubenswrapper[4967]: I1011 04:06:51.052825 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qcxk7\" (UniqueName: \"kubernetes.io/projected/292067b2-43e9-46a1-8940-08bca18ce4f0-kube-api-access-qcxk7\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:51 crc kubenswrapper[4967]: I1011 04:06:51.461993 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-87d8-account-create-9sfk2" event={"ID":"35ca69d7-b265-4ab4-a047-5fbdb349f13d","Type":"ContainerDied","Data":"7744a2457109574e0129bf874b0b6a3c1357f33b3678f5859499294749a70a0e"} Oct 11 04:06:51 crc kubenswrapper[4967]: I1011 04:06:51.462020 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-87d8-account-create-9sfk2" Oct 11 04:06:51 crc kubenswrapper[4967]: I1011 04:06:51.462028 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7744a2457109574e0129bf874b0b6a3c1357f33b3678f5859499294749a70a0e" Oct 11 04:06:51 crc kubenswrapper[4967]: I1011 04:06:51.463393 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-create-jrjv4" event={"ID":"b21fd878-e30e-498f-88f3-8a4f278ee0ec","Type":"ContainerDied","Data":"26b6cc7ecc0654b65421f95287eafebaa1864d53012043242f340eb159c8cccd"} Oct 11 04:06:51 crc kubenswrapper[4967]: I1011 04:06:51.467320 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="26b6cc7ecc0654b65421f95287eafebaa1864d53012043242f340eb159c8cccd" Oct 11 04:06:51 crc kubenswrapper[4967]: I1011 04:06:51.463403 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-create-jrjv4" Oct 11 04:06:51 crc kubenswrapper[4967]: I1011 04:06:51.475341 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-d5f2-account-create-dfps4" event={"ID":"292067b2-43e9-46a1-8940-08bca18ce4f0","Type":"ContainerDied","Data":"3301fecb776fcf18cf5bfa739aa5eb47335f435f607270d6dcc8135d38e96af6"} Oct 11 04:06:51 crc kubenswrapper[4967]: I1011 04:06:51.475418 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3301fecb776fcf18cf5bfa739aa5eb47335f435f607270d6dcc8135d38e96af6" Oct 11 04:06:51 crc kubenswrapper[4967]: I1011 04:06:51.475419 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-d5f2-account-create-dfps4" Oct 11 04:06:53 crc kubenswrapper[4967]: I1011 04:06:53.490327 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c09e3333-ad4d-495f-8973-7e2d2f0931d4-etc-swift\") pod \"swift-storage-0\" (UID: \"c09e3333-ad4d-495f-8973-7e2d2f0931d4\") " pod="openstack/swift-storage-0" Oct 11 04:06:53 crc kubenswrapper[4967]: E1011 04:06:53.490957 4967 projected.go:288] Couldn't get configMap openstack/swift-ring-files: configmap "swift-ring-files" not found Oct 11 04:06:53 crc kubenswrapper[4967]: E1011 04:06:53.490974 4967 projected.go:194] Error preparing data for projected volume etc-swift for pod openstack/swift-storage-0: configmap "swift-ring-files" not found Oct 11 04:06:53 crc kubenswrapper[4967]: E1011 04:06:53.491039 4967 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/c09e3333-ad4d-495f-8973-7e2d2f0931d4-etc-swift podName:c09e3333-ad4d-495f-8973-7e2d2f0931d4 nodeName:}" failed. No retries permitted until 2025-10-11 04:07:09.491021688 +0000 UTC m=+957.454230621 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "etc-swift" (UniqueName: "kubernetes.io/projected/c09e3333-ad4d-495f-8973-7e2d2f0931d4-etc-swift") pod "swift-storage-0" (UID: "c09e3333-ad4d-495f-8973-7e2d2f0931d4") : configmap "swift-ring-files" not found Oct 11 04:06:55 crc kubenswrapper[4967]: I1011 04:06:55.517936 4967 generic.go:334] "Generic (PLEG): container finished" podID="3f2085d6-0a42-4e1e-b52d-e9379d2a4838" containerID="0b09663baaa990bafaf876c5dd5d886e11ab30a188f37d3af2f6327d91da9462" exitCode=0 Oct 11 04:06:55 crc kubenswrapper[4967]: I1011 04:06:55.518131 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-6rr4l" event={"ID":"3f2085d6-0a42-4e1e-b52d-e9379d2a4838","Type":"ContainerDied","Data":"0b09663baaa990bafaf876c5dd5d886e11ab30a188f37d3af2f6327d91da9462"} Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.242255 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/ovn-controller-wvrnq" podUID="acad8def-db3f-40a3-a95d-28b7503d4aae" containerName="ovn-controller" probeResult="failure" output=< Oct 11 04:06:56 crc kubenswrapper[4967]: ERROR - ovn-controller connection status is 'not connected', expecting 'connected' status Oct 11 04:06:56 crc kubenswrapper[4967]: > Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.257695 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-7b5rp" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.261462 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-ovs-7b5rp" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.512240 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ovn-controller-wvrnq-config-qx7gd"] Oct 11 04:06:56 crc kubenswrapper[4967]: E1011 04:06:56.513121 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9b792fe-ab24-4587-a1f6-2ad194c9eb83" containerName="dnsmasq-dns" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.513171 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9b792fe-ab24-4587-a1f6-2ad194c9eb83" containerName="dnsmasq-dns" Oct 11 04:06:56 crc kubenswrapper[4967]: E1011 04:06:56.513228 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="292067b2-43e9-46a1-8940-08bca18ce4f0" containerName="mariadb-account-create" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.513239 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="292067b2-43e9-46a1-8940-08bca18ce4f0" containerName="mariadb-account-create" Oct 11 04:06:56 crc kubenswrapper[4967]: E1011 04:06:56.513269 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35ca69d7-b265-4ab4-a047-5fbdb349f13d" containerName="mariadb-account-create" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.513281 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="35ca69d7-b265-4ab4-a047-5fbdb349f13d" containerName="mariadb-account-create" Oct 11 04:06:56 crc kubenswrapper[4967]: E1011 04:06:56.513293 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b21fd878-e30e-498f-88f3-8a4f278ee0ec" containerName="mariadb-database-create" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.513305 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="b21fd878-e30e-498f-88f3-8a4f278ee0ec" containerName="mariadb-database-create" Oct 11 04:06:56 crc kubenswrapper[4967]: E1011 04:06:56.513362 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9b792fe-ab24-4587-a1f6-2ad194c9eb83" containerName="init" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.513373 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9b792fe-ab24-4587-a1f6-2ad194c9eb83" containerName="init" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.514330 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9b792fe-ab24-4587-a1f6-2ad194c9eb83" containerName="dnsmasq-dns" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.514389 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="35ca69d7-b265-4ab4-a047-5fbdb349f13d" containerName="mariadb-account-create" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.514439 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="292067b2-43e9-46a1-8940-08bca18ce4f0" containerName="mariadb-account-create" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.514467 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="b21fd878-e30e-498f-88f3-8a4f278ee0ec" containerName="mariadb-database-create" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.517538 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-wvrnq-config-qx7gd" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.524564 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"ovncontroller-extra-scripts" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.536400 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-wvrnq-config-qx7gd"] Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.541263 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d2713c56-4833-43cd-a956-31296bb24c4f-var-log-ovn\") pod \"ovn-controller-wvrnq-config-qx7gd\" (UID: \"d2713c56-4833-43cd-a956-31296bb24c4f\") " pod="openstack/ovn-controller-wvrnq-config-qx7gd" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.541362 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d2713c56-4833-43cd-a956-31296bb24c4f-var-run-ovn\") pod \"ovn-controller-wvrnq-config-qx7gd\" (UID: \"d2713c56-4833-43cd-a956-31296bb24c4f\") " pod="openstack/ovn-controller-wvrnq-config-qx7gd" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.541447 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d2713c56-4833-43cd-a956-31296bb24c4f-scripts\") pod \"ovn-controller-wvrnq-config-qx7gd\" (UID: \"d2713c56-4833-43cd-a956-31296bb24c4f\") " pod="openstack/ovn-controller-wvrnq-config-qx7gd" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.541509 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xjbcg\" (UniqueName: \"kubernetes.io/projected/d2713c56-4833-43cd-a956-31296bb24c4f-kube-api-access-xjbcg\") pod \"ovn-controller-wvrnq-config-qx7gd\" (UID: \"d2713c56-4833-43cd-a956-31296bb24c4f\") " pod="openstack/ovn-controller-wvrnq-config-qx7gd" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.541661 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d2713c56-4833-43cd-a956-31296bb24c4f-additional-scripts\") pod \"ovn-controller-wvrnq-config-qx7gd\" (UID: \"d2713c56-4833-43cd-a956-31296bb24c4f\") " pod="openstack/ovn-controller-wvrnq-config-qx7gd" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.541724 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d2713c56-4833-43cd-a956-31296bb24c4f-var-run\") pod \"ovn-controller-wvrnq-config-qx7gd\" (UID: \"d2713c56-4833-43cd-a956-31296bb24c4f\") " pod="openstack/ovn-controller-wvrnq-config-qx7gd" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.642776 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d2713c56-4833-43cd-a956-31296bb24c4f-var-log-ovn\") pod \"ovn-controller-wvrnq-config-qx7gd\" (UID: \"d2713c56-4833-43cd-a956-31296bb24c4f\") " pod="openstack/ovn-controller-wvrnq-config-qx7gd" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.642850 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d2713c56-4833-43cd-a956-31296bb24c4f-var-run-ovn\") pod \"ovn-controller-wvrnq-config-qx7gd\" (UID: \"d2713c56-4833-43cd-a956-31296bb24c4f\") " pod="openstack/ovn-controller-wvrnq-config-qx7gd" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.642916 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d2713c56-4833-43cd-a956-31296bb24c4f-scripts\") pod \"ovn-controller-wvrnq-config-qx7gd\" (UID: \"d2713c56-4833-43cd-a956-31296bb24c4f\") " pod="openstack/ovn-controller-wvrnq-config-qx7gd" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.642951 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xjbcg\" (UniqueName: \"kubernetes.io/projected/d2713c56-4833-43cd-a956-31296bb24c4f-kube-api-access-xjbcg\") pod \"ovn-controller-wvrnq-config-qx7gd\" (UID: \"d2713c56-4833-43cd-a956-31296bb24c4f\") " pod="openstack/ovn-controller-wvrnq-config-qx7gd" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.643009 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d2713c56-4833-43cd-a956-31296bb24c4f-additional-scripts\") pod \"ovn-controller-wvrnq-config-qx7gd\" (UID: \"d2713c56-4833-43cd-a956-31296bb24c4f\") " pod="openstack/ovn-controller-wvrnq-config-qx7gd" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.643032 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d2713c56-4833-43cd-a956-31296bb24c4f-var-run\") pod \"ovn-controller-wvrnq-config-qx7gd\" (UID: \"d2713c56-4833-43cd-a956-31296bb24c4f\") " pod="openstack/ovn-controller-wvrnq-config-qx7gd" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.643153 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d2713c56-4833-43cd-a956-31296bb24c4f-var-run-ovn\") pod \"ovn-controller-wvrnq-config-qx7gd\" (UID: \"d2713c56-4833-43cd-a956-31296bb24c4f\") " pod="openstack/ovn-controller-wvrnq-config-qx7gd" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.643152 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d2713c56-4833-43cd-a956-31296bb24c4f-var-log-ovn\") pod \"ovn-controller-wvrnq-config-qx7gd\" (UID: \"d2713c56-4833-43cd-a956-31296bb24c4f\") " pod="openstack/ovn-controller-wvrnq-config-qx7gd" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.643245 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d2713c56-4833-43cd-a956-31296bb24c4f-var-run\") pod \"ovn-controller-wvrnq-config-qx7gd\" (UID: \"d2713c56-4833-43cd-a956-31296bb24c4f\") " pod="openstack/ovn-controller-wvrnq-config-qx7gd" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.643998 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d2713c56-4833-43cd-a956-31296bb24c4f-additional-scripts\") pod \"ovn-controller-wvrnq-config-qx7gd\" (UID: \"d2713c56-4833-43cd-a956-31296bb24c4f\") " pod="openstack/ovn-controller-wvrnq-config-qx7gd" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.644982 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d2713c56-4833-43cd-a956-31296bb24c4f-scripts\") pod \"ovn-controller-wvrnq-config-qx7gd\" (UID: \"d2713c56-4833-43cd-a956-31296bb24c4f\") " pod="openstack/ovn-controller-wvrnq-config-qx7gd" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.679572 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xjbcg\" (UniqueName: \"kubernetes.io/projected/d2713c56-4833-43cd-a956-31296bb24c4f-kube-api-access-xjbcg\") pod \"ovn-controller-wvrnq-config-qx7gd\" (UID: \"d2713c56-4833-43cd-a956-31296bb24c4f\") " pod="openstack/ovn-controller-wvrnq-config-qx7gd" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.854420 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-wvrnq-config-qx7gd" Oct 11 04:06:56 crc kubenswrapper[4967]: I1011 04:06:56.969685 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6rr4l" Oct 11 04:06:57 crc kubenswrapper[4967]: I1011 04:06:57.049431 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l95j8\" (UniqueName: \"kubernetes.io/projected/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-kube-api-access-l95j8\") pod \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\" (UID: \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\") " Oct 11 04:06:57 crc kubenswrapper[4967]: I1011 04:06:57.049478 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-swiftconf\") pod \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\" (UID: \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\") " Oct 11 04:06:57 crc kubenswrapper[4967]: I1011 04:06:57.049502 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-dispersionconf\") pod \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\" (UID: \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\") " Oct 11 04:06:57 crc kubenswrapper[4967]: I1011 04:06:57.049552 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-etc-swift\") pod \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\" (UID: \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\") " Oct 11 04:06:57 crc kubenswrapper[4967]: I1011 04:06:57.049601 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-ring-data-devices\") pod \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\" (UID: \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\") " Oct 11 04:06:57 crc kubenswrapper[4967]: I1011 04:06:57.049689 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-combined-ca-bundle\") pod \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\" (UID: \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\") " Oct 11 04:06:57 crc kubenswrapper[4967]: I1011 04:06:57.050312 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-ring-data-devices" (OuterVolumeSpecName: "ring-data-devices") pod "3f2085d6-0a42-4e1e-b52d-e9379d2a4838" (UID: "3f2085d6-0a42-4e1e-b52d-e9379d2a4838"). InnerVolumeSpecName "ring-data-devices". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:06:57 crc kubenswrapper[4967]: I1011 04:06:57.050462 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-etc-swift" (OuterVolumeSpecName: "etc-swift") pod "3f2085d6-0a42-4e1e-b52d-e9379d2a4838" (UID: "3f2085d6-0a42-4e1e-b52d-e9379d2a4838"). InnerVolumeSpecName "etc-swift". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:06:57 crc kubenswrapper[4967]: I1011 04:06:57.050504 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-scripts\") pod \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\" (UID: \"3f2085d6-0a42-4e1e-b52d-e9379d2a4838\") " Oct 11 04:06:57 crc kubenswrapper[4967]: I1011 04:06:57.051028 4967 reconciler_common.go:293] "Volume detached for volume \"etc-swift\" (UniqueName: \"kubernetes.io/empty-dir/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-etc-swift\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:57 crc kubenswrapper[4967]: I1011 04:06:57.051044 4967 reconciler_common.go:293] "Volume detached for volume \"ring-data-devices\" (UniqueName: \"kubernetes.io/configmap/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-ring-data-devices\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:57 crc kubenswrapper[4967]: I1011 04:06:57.054378 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-kube-api-access-l95j8" (OuterVolumeSpecName: "kube-api-access-l95j8") pod "3f2085d6-0a42-4e1e-b52d-e9379d2a4838" (UID: "3f2085d6-0a42-4e1e-b52d-e9379d2a4838"). InnerVolumeSpecName "kube-api-access-l95j8". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:06:57 crc kubenswrapper[4967]: I1011 04:06:57.070324 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-dispersionconf" (OuterVolumeSpecName: "dispersionconf") pod "3f2085d6-0a42-4e1e-b52d-e9379d2a4838" (UID: "3f2085d6-0a42-4e1e-b52d-e9379d2a4838"). InnerVolumeSpecName "dispersionconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:06:57 crc kubenswrapper[4967]: I1011 04:06:57.070808 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-scripts" (OuterVolumeSpecName: "scripts") pod "3f2085d6-0a42-4e1e-b52d-e9379d2a4838" (UID: "3f2085d6-0a42-4e1e-b52d-e9379d2a4838"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:06:57 crc kubenswrapper[4967]: I1011 04:06:57.071871 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3f2085d6-0a42-4e1e-b52d-e9379d2a4838" (UID: "3f2085d6-0a42-4e1e-b52d-e9379d2a4838"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:06:57 crc kubenswrapper[4967]: I1011 04:06:57.079298 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-swiftconf" (OuterVolumeSpecName: "swiftconf") pod "3f2085d6-0a42-4e1e-b52d-e9379d2a4838" (UID: "3f2085d6-0a42-4e1e-b52d-e9379d2a4838"). InnerVolumeSpecName "swiftconf". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:06:57 crc kubenswrapper[4967]: I1011 04:06:57.153583 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:57 crc kubenswrapper[4967]: I1011 04:06:57.153618 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:57 crc kubenswrapper[4967]: I1011 04:06:57.153629 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l95j8\" (UniqueName: \"kubernetes.io/projected/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-kube-api-access-l95j8\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:57 crc kubenswrapper[4967]: I1011 04:06:57.153639 4967 reconciler_common.go:293] "Volume detached for volume \"swiftconf\" (UniqueName: \"kubernetes.io/secret/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-swiftconf\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:57 crc kubenswrapper[4967]: I1011 04:06:57.153647 4967 reconciler_common.go:293] "Volume detached for volume \"dispersionconf\" (UniqueName: \"kubernetes.io/secret/3f2085d6-0a42-4e1e-b52d-e9379d2a4838-dispersionconf\") on node \"crc\" DevicePath \"\"" Oct 11 04:06:57 crc kubenswrapper[4967]: I1011 04:06:57.297019 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ovn-controller-wvrnq-config-qx7gd"] Oct 11 04:06:57 crc kubenswrapper[4967]: I1011 04:06:57.541019 4967 generic.go:334] "Generic (PLEG): container finished" podID="3d923c19-a45c-4cb7-9f19-e770c5fe66ed" containerID="90a1eab201a7f4eef8d6e63d4754f3e37c2c7b5857764dda95a5217b63654438" exitCode=0 Oct 11 04:06:57 crc kubenswrapper[4967]: I1011 04:06:57.541309 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3d923c19-a45c-4cb7-9f19-e770c5fe66ed","Type":"ContainerDied","Data":"90a1eab201a7f4eef8d6e63d4754f3e37c2c7b5857764dda95a5217b63654438"} Oct 11 04:06:57 crc kubenswrapper[4967]: I1011 04:06:57.545560 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-ring-rebalance-6rr4l" event={"ID":"3f2085d6-0a42-4e1e-b52d-e9379d2a4838","Type":"ContainerDied","Data":"08a13f6849a09fdd5a5a4f60260b83af9b67e0798d1e5393e3a69d13c5b3d8f1"} Oct 11 04:06:57 crc kubenswrapper[4967]: I1011 04:06:57.545581 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/swift-ring-rebalance-6rr4l" Oct 11 04:06:57 crc kubenswrapper[4967]: I1011 04:06:57.545595 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="08a13f6849a09fdd5a5a4f60260b83af9b67e0798d1e5393e3a69d13c5b3d8f1" Oct 11 04:06:57 crc kubenswrapper[4967]: I1011 04:06:57.547780 4967 generic.go:334] "Generic (PLEG): container finished" podID="0c315653-59d1-429f-be9c-c1ac826cff49" containerID="55e8d168f569f266bd42f0f73e79f782f1b67630b934f6721fb999704f507332" exitCode=0 Oct 11 04:06:57 crc kubenswrapper[4967]: I1011 04:06:57.547827 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0c315653-59d1-429f-be9c-c1ac826cff49","Type":"ContainerDied","Data":"55e8d168f569f266bd42f0f73e79f782f1b67630b934f6721fb999704f507332"} Oct 11 04:06:57 crc kubenswrapper[4967]: I1011 04:06:57.553561 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-wvrnq-config-qx7gd" event={"ID":"d2713c56-4833-43cd-a956-31296bb24c4f","Type":"ContainerStarted","Data":"2b4697304961680253f8301d1bc7841dcac1179f3a40e0d2e038bd4d00fbc381"} Oct 11 04:06:58 crc kubenswrapper[4967]: I1011 04:06:58.562499 4967 generic.go:334] "Generic (PLEG): container finished" podID="d2713c56-4833-43cd-a956-31296bb24c4f" containerID="f8dc539c1c9ddd669160c80b33dbd71e7b11e50401ace2b49135c047d6adf682" exitCode=0 Oct 11 04:06:58 crc kubenswrapper[4967]: I1011 04:06:58.562571 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-wvrnq-config-qx7gd" event={"ID":"d2713c56-4833-43cd-a956-31296bb24c4f","Type":"ContainerDied","Data":"f8dc539c1c9ddd669160c80b33dbd71e7b11e50401ace2b49135c047d6adf682"} Oct 11 04:06:58 crc kubenswrapper[4967]: I1011 04:06:58.567318 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3d923c19-a45c-4cb7-9f19-e770c5fe66ed","Type":"ContainerStarted","Data":"3a0e9dd337a635129967491448fa27c66efed6b3f2cf226b43e39c2a2861a55c"} Oct 11 04:06:58 crc kubenswrapper[4967]: I1011 04:06:58.567636 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 11 04:06:58 crc kubenswrapper[4967]: I1011 04:06:58.569402 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0c315653-59d1-429f-be9c-c1ac826cff49","Type":"ContainerStarted","Data":"6954a528ee16848ee7e858a4e86aaad17c994e0a8493821729091c842af849a0"} Oct 11 04:06:58 crc kubenswrapper[4967]: I1011 04:06:58.569607 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:06:58 crc kubenswrapper[4967]: I1011 04:06:58.637593 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=50.023063501 podStartE2EDuration="59.637579868s" podCreationTimestamp="2025-10-11 04:05:59 +0000 UTC" firstStartedPulling="2025-10-11 04:06:12.704737672 +0000 UTC m=+900.667946605" lastFinishedPulling="2025-10-11 04:06:22.319254039 +0000 UTC m=+910.282462972" observedRunningTime="2025-10-11 04:06:58.633870141 +0000 UTC m=+946.597079074" watchObservedRunningTime="2025-10-11 04:06:58.637579868 +0000 UTC m=+946.600788791" Oct 11 04:06:58 crc kubenswrapper[4967]: I1011 04:06:58.638746 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=49.22008605 podStartE2EDuration="59.638741145s" podCreationTimestamp="2025-10-11 04:05:59 +0000 UTC" firstStartedPulling="2025-10-11 04:06:12.640023503 +0000 UTC m=+900.603232436" lastFinishedPulling="2025-10-11 04:06:23.058678598 +0000 UTC m=+911.021887531" observedRunningTime="2025-10-11 04:06:58.612586018 +0000 UTC m=+946.575794971" watchObservedRunningTime="2025-10-11 04:06:58.638741145 +0000 UTC m=+946.601950068" Oct 11 04:06:59 crc kubenswrapper[4967]: I1011 04:06:59.957594 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-7892-account-create-2qg5w"] Oct 11 04:06:59 crc kubenswrapper[4967]: E1011 04:06:59.958289 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3f2085d6-0a42-4e1e-b52d-e9379d2a4838" containerName="swift-ring-rebalance" Oct 11 04:06:59 crc kubenswrapper[4967]: I1011 04:06:59.958303 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="3f2085d6-0a42-4e1e-b52d-e9379d2a4838" containerName="swift-ring-rebalance" Oct 11 04:06:59 crc kubenswrapper[4967]: I1011 04:06:59.958587 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="3f2085d6-0a42-4e1e-b52d-e9379d2a4838" containerName="swift-ring-rebalance" Oct 11 04:06:59 crc kubenswrapper[4967]: I1011 04:06:59.959229 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-7892-account-create-2qg5w" Oct 11 04:06:59 crc kubenswrapper[4967]: I1011 04:06:59.961271 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-db-secret" Oct 11 04:06:59 crc kubenswrapper[4967]: I1011 04:06:59.968098 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-7892-account-create-2qg5w"] Oct 11 04:06:59 crc kubenswrapper[4967]: I1011 04:06:59.976860 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-wvrnq-config-qx7gd" Oct 11 04:07:00 crc kubenswrapper[4967]: I1011 04:07:00.104102 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d2713c56-4833-43cd-a956-31296bb24c4f-additional-scripts\") pod \"d2713c56-4833-43cd-a956-31296bb24c4f\" (UID: \"d2713c56-4833-43cd-a956-31296bb24c4f\") " Oct 11 04:07:00 crc kubenswrapper[4967]: I1011 04:07:00.104320 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d2713c56-4833-43cd-a956-31296bb24c4f-scripts\") pod \"d2713c56-4833-43cd-a956-31296bb24c4f\" (UID: \"d2713c56-4833-43cd-a956-31296bb24c4f\") " Oct 11 04:07:00 crc kubenswrapper[4967]: I1011 04:07:00.104991 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2713c56-4833-43cd-a956-31296bb24c4f-additional-scripts" (OuterVolumeSpecName: "additional-scripts") pod "d2713c56-4833-43cd-a956-31296bb24c4f" (UID: "d2713c56-4833-43cd-a956-31296bb24c4f"). InnerVolumeSpecName "additional-scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:07:00 crc kubenswrapper[4967]: I1011 04:07:00.105486 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/d2713c56-4833-43cd-a956-31296bb24c4f-scripts" (OuterVolumeSpecName: "scripts") pod "d2713c56-4833-43cd-a956-31296bb24c4f" (UID: "d2713c56-4833-43cd-a956-31296bb24c4f"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:07:00 crc kubenswrapper[4967]: I1011 04:07:00.105607 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d2713c56-4833-43cd-a956-31296bb24c4f-var-run\") pod \"d2713c56-4833-43cd-a956-31296bb24c4f\" (UID: \"d2713c56-4833-43cd-a956-31296bb24c4f\") " Oct 11 04:07:00 crc kubenswrapper[4967]: I1011 04:07:00.105677 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d2713c56-4833-43cd-a956-31296bb24c4f-var-run" (OuterVolumeSpecName: "var-run") pod "d2713c56-4833-43cd-a956-31296bb24c4f" (UID: "d2713c56-4833-43cd-a956-31296bb24c4f"). InnerVolumeSpecName "var-run". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 04:07:00 crc kubenswrapper[4967]: I1011 04:07:00.105640 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xjbcg\" (UniqueName: \"kubernetes.io/projected/d2713c56-4833-43cd-a956-31296bb24c4f-kube-api-access-xjbcg\") pod \"d2713c56-4833-43cd-a956-31296bb24c4f\" (UID: \"d2713c56-4833-43cd-a956-31296bb24c4f\") " Oct 11 04:07:00 crc kubenswrapper[4967]: I1011 04:07:00.105792 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d2713c56-4833-43cd-a956-31296bb24c4f-var-run-ovn\") pod \"d2713c56-4833-43cd-a956-31296bb24c4f\" (UID: \"d2713c56-4833-43cd-a956-31296bb24c4f\") " Oct 11 04:07:00 crc kubenswrapper[4967]: I1011 04:07:00.105863 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d2713c56-4833-43cd-a956-31296bb24c4f-var-run-ovn" (OuterVolumeSpecName: "var-run-ovn") pod "d2713c56-4833-43cd-a956-31296bb24c4f" (UID: "d2713c56-4833-43cd-a956-31296bb24c4f"). InnerVolumeSpecName "var-run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 04:07:00 crc kubenswrapper[4967]: I1011 04:07:00.105934 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d2713c56-4833-43cd-a956-31296bb24c4f-var-log-ovn\") pod \"d2713c56-4833-43cd-a956-31296bb24c4f\" (UID: \"d2713c56-4833-43cd-a956-31296bb24c4f\") " Oct 11 04:07:00 crc kubenswrapper[4967]: I1011 04:07:00.106009 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/d2713c56-4833-43cd-a956-31296bb24c4f-var-log-ovn" (OuterVolumeSpecName: "var-log-ovn") pod "d2713c56-4833-43cd-a956-31296bb24c4f" (UID: "d2713c56-4833-43cd-a956-31296bb24c4f"). InnerVolumeSpecName "var-log-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 04:07:00 crc kubenswrapper[4967]: I1011 04:07:00.106414 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvqhk\" (UniqueName: \"kubernetes.io/projected/3af91e9f-81cd-4cdb-89d2-a1c50f55f62a-kube-api-access-mvqhk\") pod \"glance-7892-account-create-2qg5w\" (UID: \"3af91e9f-81cd-4cdb-89d2-a1c50f55f62a\") " pod="openstack/glance-7892-account-create-2qg5w" Oct 11 04:07:00 crc kubenswrapper[4967]: I1011 04:07:00.106523 4967 reconciler_common.go:293] "Volume detached for volume \"var-run-ovn\" (UniqueName: \"kubernetes.io/host-path/d2713c56-4833-43cd-a956-31296bb24c4f-var-run-ovn\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:00 crc kubenswrapper[4967]: I1011 04:07:00.106546 4967 reconciler_common.go:293] "Volume detached for volume \"var-log-ovn\" (UniqueName: \"kubernetes.io/host-path/d2713c56-4833-43cd-a956-31296bb24c4f-var-log-ovn\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:00 crc kubenswrapper[4967]: I1011 04:07:00.106560 4967 reconciler_common.go:293] "Volume detached for volume \"additional-scripts\" (UniqueName: \"kubernetes.io/configmap/d2713c56-4833-43cd-a956-31296bb24c4f-additional-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:00 crc kubenswrapper[4967]: I1011 04:07:00.106572 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/configmap/d2713c56-4833-43cd-a956-31296bb24c4f-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:00 crc kubenswrapper[4967]: I1011 04:07:00.106583 4967 reconciler_common.go:293] "Volume detached for volume \"var-run\" (UniqueName: \"kubernetes.io/host-path/d2713c56-4833-43cd-a956-31296bb24c4f-var-run\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:00 crc kubenswrapper[4967]: I1011 04:07:00.110917 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d2713c56-4833-43cd-a956-31296bb24c4f-kube-api-access-xjbcg" (OuterVolumeSpecName: "kube-api-access-xjbcg") pod "d2713c56-4833-43cd-a956-31296bb24c4f" (UID: "d2713c56-4833-43cd-a956-31296bb24c4f"). InnerVolumeSpecName "kube-api-access-xjbcg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:07:00 crc kubenswrapper[4967]: I1011 04:07:00.207952 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvqhk\" (UniqueName: \"kubernetes.io/projected/3af91e9f-81cd-4cdb-89d2-a1c50f55f62a-kube-api-access-mvqhk\") pod \"glance-7892-account-create-2qg5w\" (UID: \"3af91e9f-81cd-4cdb-89d2-a1c50f55f62a\") " pod="openstack/glance-7892-account-create-2qg5w" Oct 11 04:07:00 crc kubenswrapper[4967]: I1011 04:07:00.208103 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xjbcg\" (UniqueName: \"kubernetes.io/projected/d2713c56-4833-43cd-a956-31296bb24c4f-kube-api-access-xjbcg\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:00 crc kubenswrapper[4967]: I1011 04:07:00.228995 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvqhk\" (UniqueName: \"kubernetes.io/projected/3af91e9f-81cd-4cdb-89d2-a1c50f55f62a-kube-api-access-mvqhk\") pod \"glance-7892-account-create-2qg5w\" (UID: \"3af91e9f-81cd-4cdb-89d2-a1c50f55f62a\") " pod="openstack/glance-7892-account-create-2qg5w" Oct 11 04:07:00 crc kubenswrapper[4967]: I1011 04:07:00.288529 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-7892-account-create-2qg5w" Oct 11 04:07:00 crc kubenswrapper[4967]: I1011 04:07:00.572410 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-7892-account-create-2qg5w"] Oct 11 04:07:00 crc kubenswrapper[4967]: I1011 04:07:00.586799 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ovn-controller-wvrnq-config-qx7gd" event={"ID":"d2713c56-4833-43cd-a956-31296bb24c4f","Type":"ContainerDied","Data":"2b4697304961680253f8301d1bc7841dcac1179f3a40e0d2e038bd4d00fbc381"} Oct 11 04:07:00 crc kubenswrapper[4967]: I1011 04:07:00.586832 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2b4697304961680253f8301d1bc7841dcac1179f3a40e0d2e038bd4d00fbc381" Oct 11 04:07:00 crc kubenswrapper[4967]: I1011 04:07:00.586864 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ovn-controller-wvrnq-config-qx7gd" Oct 11 04:07:01 crc kubenswrapper[4967]: I1011 04:07:01.110064 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ovn-controller-wvrnq-config-qx7gd"] Oct 11 04:07:01 crc kubenswrapper[4967]: I1011 04:07:01.117095 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ovn-controller-wvrnq-config-qx7gd"] Oct 11 04:07:01 crc kubenswrapper[4967]: I1011 04:07:01.235024 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ovn-controller-wvrnq" Oct 11 04:07:01 crc kubenswrapper[4967]: I1011 04:07:01.594890 4967 generic.go:334] "Generic (PLEG): container finished" podID="3af91e9f-81cd-4cdb-89d2-a1c50f55f62a" containerID="0fcbd3c9399c32b7d84f732ff8c9b109eb4a2f153f4c9cd8cc2852bb1f1e5fbe" exitCode=0 Oct 11 04:07:01 crc kubenswrapper[4967]: I1011 04:07:01.594935 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-7892-account-create-2qg5w" event={"ID":"3af91e9f-81cd-4cdb-89d2-a1c50f55f62a","Type":"ContainerDied","Data":"0fcbd3c9399c32b7d84f732ff8c9b109eb4a2f153f4c9cd8cc2852bb1f1e5fbe"} Oct 11 04:07:01 crc kubenswrapper[4967]: I1011 04:07:01.594965 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-7892-account-create-2qg5w" event={"ID":"3af91e9f-81cd-4cdb-89d2-a1c50f55f62a","Type":"ContainerStarted","Data":"c77d8d8d21648b5642dc5cdf09511fdc6c382c8981328acb22019f332d589e9a"} Oct 11 04:07:02 crc kubenswrapper[4967]: I1011 04:07:02.825321 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d2713c56-4833-43cd-a956-31296bb24c4f" path="/var/lib/kubelet/pods/d2713c56-4833-43cd-a956-31296bb24c4f/volumes" Oct 11 04:07:02 crc kubenswrapper[4967]: I1011 04:07:02.922137 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-7892-account-create-2qg5w" Oct 11 04:07:03 crc kubenswrapper[4967]: I1011 04:07:03.052689 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mvqhk\" (UniqueName: \"kubernetes.io/projected/3af91e9f-81cd-4cdb-89d2-a1c50f55f62a-kube-api-access-mvqhk\") pod \"3af91e9f-81cd-4cdb-89d2-a1c50f55f62a\" (UID: \"3af91e9f-81cd-4cdb-89d2-a1c50f55f62a\") " Oct 11 04:07:03 crc kubenswrapper[4967]: I1011 04:07:03.060431 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3af91e9f-81cd-4cdb-89d2-a1c50f55f62a-kube-api-access-mvqhk" (OuterVolumeSpecName: "kube-api-access-mvqhk") pod "3af91e9f-81cd-4cdb-89d2-a1c50f55f62a" (UID: "3af91e9f-81cd-4cdb-89d2-a1c50f55f62a"). InnerVolumeSpecName "kube-api-access-mvqhk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:07:03 crc kubenswrapper[4967]: I1011 04:07:03.155931 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mvqhk\" (UniqueName: \"kubernetes.io/projected/3af91e9f-81cd-4cdb-89d2-a1c50f55f62a-kube-api-access-mvqhk\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:03 crc kubenswrapper[4967]: I1011 04:07:03.613862 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-7892-account-create-2qg5w" event={"ID":"3af91e9f-81cd-4cdb-89d2-a1c50f55f62a","Type":"ContainerDied","Data":"c77d8d8d21648b5642dc5cdf09511fdc6c382c8981328acb22019f332d589e9a"} Oct 11 04:07:03 crc kubenswrapper[4967]: I1011 04:07:03.613916 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c77d8d8d21648b5642dc5cdf09511fdc6c382c8981328acb22019f332d589e9a" Oct 11 04:07:03 crc kubenswrapper[4967]: I1011 04:07:03.614343 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-7892-account-create-2qg5w" Oct 11 04:07:05 crc kubenswrapper[4967]: I1011 04:07:05.197724 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-db-sync-mgcb9"] Oct 11 04:07:05 crc kubenswrapper[4967]: E1011 04:07:05.198527 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d2713c56-4833-43cd-a956-31296bb24c4f" containerName="ovn-config" Oct 11 04:07:05 crc kubenswrapper[4967]: I1011 04:07:05.198548 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="d2713c56-4833-43cd-a956-31296bb24c4f" containerName="ovn-config" Oct 11 04:07:05 crc kubenswrapper[4967]: E1011 04:07:05.198579 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3af91e9f-81cd-4cdb-89d2-a1c50f55f62a" containerName="mariadb-account-create" Oct 11 04:07:05 crc kubenswrapper[4967]: I1011 04:07:05.198596 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="3af91e9f-81cd-4cdb-89d2-a1c50f55f62a" containerName="mariadb-account-create" Oct 11 04:07:05 crc kubenswrapper[4967]: I1011 04:07:05.198956 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="d2713c56-4833-43cd-a956-31296bb24c4f" containerName="ovn-config" Oct 11 04:07:05 crc kubenswrapper[4967]: I1011 04:07:05.198985 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="3af91e9f-81cd-4cdb-89d2-a1c50f55f62a" containerName="mariadb-account-create" Oct 11 04:07:05 crc kubenswrapper[4967]: I1011 04:07:05.199846 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-mgcb9" Oct 11 04:07:05 crc kubenswrapper[4967]: I1011 04:07:05.203656 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-config-data" Oct 11 04:07:05 crc kubenswrapper[4967]: I1011 04:07:05.203666 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-2d8h2" Oct 11 04:07:05 crc kubenswrapper[4967]: I1011 04:07:05.214937 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-mgcb9"] Oct 11 04:07:05 crc kubenswrapper[4967]: I1011 04:07:05.291910 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jkbwn\" (UniqueName: \"kubernetes.io/projected/df08696a-e0e9-4d45-9ffb-c41eff1b5fb4-kube-api-access-jkbwn\") pod \"glance-db-sync-mgcb9\" (UID: \"df08696a-e0e9-4d45-9ffb-c41eff1b5fb4\") " pod="openstack/glance-db-sync-mgcb9" Oct 11 04:07:05 crc kubenswrapper[4967]: I1011 04:07:05.294804 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df08696a-e0e9-4d45-9ffb-c41eff1b5fb4-combined-ca-bundle\") pod \"glance-db-sync-mgcb9\" (UID: \"df08696a-e0e9-4d45-9ffb-c41eff1b5fb4\") " pod="openstack/glance-db-sync-mgcb9" Oct 11 04:07:05 crc kubenswrapper[4967]: I1011 04:07:05.295017 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df08696a-e0e9-4d45-9ffb-c41eff1b5fb4-config-data\") pod \"glance-db-sync-mgcb9\" (UID: \"df08696a-e0e9-4d45-9ffb-c41eff1b5fb4\") " pod="openstack/glance-db-sync-mgcb9" Oct 11 04:07:05 crc kubenswrapper[4967]: I1011 04:07:05.295098 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/df08696a-e0e9-4d45-9ffb-c41eff1b5fb4-db-sync-config-data\") pod \"glance-db-sync-mgcb9\" (UID: \"df08696a-e0e9-4d45-9ffb-c41eff1b5fb4\") " pod="openstack/glance-db-sync-mgcb9" Oct 11 04:07:05 crc kubenswrapper[4967]: I1011 04:07:05.396093 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jkbwn\" (UniqueName: \"kubernetes.io/projected/df08696a-e0e9-4d45-9ffb-c41eff1b5fb4-kube-api-access-jkbwn\") pod \"glance-db-sync-mgcb9\" (UID: \"df08696a-e0e9-4d45-9ffb-c41eff1b5fb4\") " pod="openstack/glance-db-sync-mgcb9" Oct 11 04:07:05 crc kubenswrapper[4967]: I1011 04:07:05.396208 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df08696a-e0e9-4d45-9ffb-c41eff1b5fb4-combined-ca-bundle\") pod \"glance-db-sync-mgcb9\" (UID: \"df08696a-e0e9-4d45-9ffb-c41eff1b5fb4\") " pod="openstack/glance-db-sync-mgcb9" Oct 11 04:07:05 crc kubenswrapper[4967]: I1011 04:07:05.396241 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df08696a-e0e9-4d45-9ffb-c41eff1b5fb4-config-data\") pod \"glance-db-sync-mgcb9\" (UID: \"df08696a-e0e9-4d45-9ffb-c41eff1b5fb4\") " pod="openstack/glance-db-sync-mgcb9" Oct 11 04:07:05 crc kubenswrapper[4967]: I1011 04:07:05.396264 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/df08696a-e0e9-4d45-9ffb-c41eff1b5fb4-db-sync-config-data\") pod \"glance-db-sync-mgcb9\" (UID: \"df08696a-e0e9-4d45-9ffb-c41eff1b5fb4\") " pod="openstack/glance-db-sync-mgcb9" Oct 11 04:07:05 crc kubenswrapper[4967]: I1011 04:07:05.400895 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/df08696a-e0e9-4d45-9ffb-c41eff1b5fb4-db-sync-config-data\") pod \"glance-db-sync-mgcb9\" (UID: \"df08696a-e0e9-4d45-9ffb-c41eff1b5fb4\") " pod="openstack/glance-db-sync-mgcb9" Oct 11 04:07:05 crc kubenswrapper[4967]: I1011 04:07:05.401388 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df08696a-e0e9-4d45-9ffb-c41eff1b5fb4-combined-ca-bundle\") pod \"glance-db-sync-mgcb9\" (UID: \"df08696a-e0e9-4d45-9ffb-c41eff1b5fb4\") " pod="openstack/glance-db-sync-mgcb9" Oct 11 04:07:05 crc kubenswrapper[4967]: I1011 04:07:05.405869 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df08696a-e0e9-4d45-9ffb-c41eff1b5fb4-config-data\") pod \"glance-db-sync-mgcb9\" (UID: \"df08696a-e0e9-4d45-9ffb-c41eff1b5fb4\") " pod="openstack/glance-db-sync-mgcb9" Oct 11 04:07:05 crc kubenswrapper[4967]: I1011 04:07:05.415202 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jkbwn\" (UniqueName: \"kubernetes.io/projected/df08696a-e0e9-4d45-9ffb-c41eff1b5fb4-kube-api-access-jkbwn\") pod \"glance-db-sync-mgcb9\" (UID: \"df08696a-e0e9-4d45-9ffb-c41eff1b5fb4\") " pod="openstack/glance-db-sync-mgcb9" Oct 11 04:07:05 crc kubenswrapper[4967]: I1011 04:07:05.565500 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-mgcb9" Oct 11 04:07:06 crc kubenswrapper[4967]: I1011 04:07:06.099787 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-db-sync-mgcb9"] Oct 11 04:07:06 crc kubenswrapper[4967]: W1011 04:07:06.107722 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddf08696a_e0e9_4d45_9ffb_c41eff1b5fb4.slice/crio-9f149671ac7f0a791e3adb6f8dd3664bd27ad1cc13b8f0e843410cb1225c5201 WatchSource:0}: Error finding container 9f149671ac7f0a791e3adb6f8dd3664bd27ad1cc13b8f0e843410cb1225c5201: Status 404 returned error can't find the container with id 9f149671ac7f0a791e3adb6f8dd3664bd27ad1cc13b8f0e843410cb1225c5201 Oct 11 04:07:06 crc kubenswrapper[4967]: I1011 04:07:06.643482 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-mgcb9" event={"ID":"df08696a-e0e9-4d45-9ffb-c41eff1b5fb4","Type":"ContainerStarted","Data":"9f149671ac7f0a791e3adb6f8dd3664bd27ad1cc13b8f0e843410cb1225c5201"} Oct 11 04:07:09 crc kubenswrapper[4967]: I1011 04:07:09.565476 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c09e3333-ad4d-495f-8973-7e2d2f0931d4-etc-swift\") pod \"swift-storage-0\" (UID: \"c09e3333-ad4d-495f-8973-7e2d2f0931d4\") " pod="openstack/swift-storage-0" Oct 11 04:07:09 crc kubenswrapper[4967]: I1011 04:07:09.573168 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/c09e3333-ad4d-495f-8973-7e2d2f0931d4-etc-swift\") pod \"swift-storage-0\" (UID: \"c09e3333-ad4d-495f-8973-7e2d2f0931d4\") " pod="openstack/swift-storage-0" Oct 11 04:07:09 crc kubenswrapper[4967]: I1011 04:07:09.756421 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-storage-0" Oct 11 04:07:10 crc kubenswrapper[4967]: I1011 04:07:10.924252 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:07:11 crc kubenswrapper[4967]: I1011 04:07:11.249241 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 11 04:07:12 crc kubenswrapper[4967]: I1011 04:07:12.496649 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-create-f5bhd"] Oct 11 04:07:12 crc kubenswrapper[4967]: I1011 04:07:12.497791 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-f5bhd" Oct 11 04:07:12 crc kubenswrapper[4967]: I1011 04:07:12.506775 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-f5bhd"] Oct 11 04:07:12 crc kubenswrapper[4967]: I1011 04:07:12.601050 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-create-257lm"] Oct 11 04:07:12 crc kubenswrapper[4967]: I1011 04:07:12.602059 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-257lm" Oct 11 04:07:12 crc kubenswrapper[4967]: I1011 04:07:12.608678 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-257lm"] Oct 11 04:07:12 crc kubenswrapper[4967]: I1011 04:07:12.637453 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r86rl\" (UniqueName: \"kubernetes.io/projected/6f16161f-b863-46ca-b2f1-7a6248f27c4d-kube-api-access-r86rl\") pod \"barbican-db-create-f5bhd\" (UID: \"6f16161f-b863-46ca-b2f1-7a6248f27c4d\") " pod="openstack/barbican-db-create-f5bhd" Oct 11 04:07:12 crc kubenswrapper[4967]: I1011 04:07:12.739592 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tc8b9\" (UniqueName: \"kubernetes.io/projected/fdb459dd-26c4-49c4-bf1f-dad807314e3d-kube-api-access-tc8b9\") pod \"cinder-db-create-257lm\" (UID: \"fdb459dd-26c4-49c4-bf1f-dad807314e3d\") " pod="openstack/cinder-db-create-257lm" Oct 11 04:07:12 crc kubenswrapper[4967]: I1011 04:07:12.739770 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r86rl\" (UniqueName: \"kubernetes.io/projected/6f16161f-b863-46ca-b2f1-7a6248f27c4d-kube-api-access-r86rl\") pod \"barbican-db-create-f5bhd\" (UID: \"6f16161f-b863-46ca-b2f1-7a6248f27c4d\") " pod="openstack/barbican-db-create-f5bhd" Oct 11 04:07:12 crc kubenswrapper[4967]: I1011 04:07:12.760539 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r86rl\" (UniqueName: \"kubernetes.io/projected/6f16161f-b863-46ca-b2f1-7a6248f27c4d-kube-api-access-r86rl\") pod \"barbican-db-create-f5bhd\" (UID: \"6f16161f-b863-46ca-b2f1-7a6248f27c4d\") " pod="openstack/barbican-db-create-f5bhd" Oct 11 04:07:12 crc kubenswrapper[4967]: I1011 04:07:12.808150 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-create-9p6rn"] Oct 11 04:07:12 crc kubenswrapper[4967]: I1011 04:07:12.809119 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9p6rn" Oct 11 04:07:12 crc kubenswrapper[4967]: I1011 04:07:12.840716 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tc8b9\" (UniqueName: \"kubernetes.io/projected/fdb459dd-26c4-49c4-bf1f-dad807314e3d-kube-api-access-tc8b9\") pod \"cinder-db-create-257lm\" (UID: \"fdb459dd-26c4-49c4-bf1f-dad807314e3d\") " pod="openstack/cinder-db-create-257lm" Oct 11 04:07:12 crc kubenswrapper[4967]: I1011 04:07:12.841849 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-9p6rn"] Oct 11 04:07:12 crc kubenswrapper[4967]: I1011 04:07:12.860332 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tc8b9\" (UniqueName: \"kubernetes.io/projected/fdb459dd-26c4-49c4-bf1f-dad807314e3d-kube-api-access-tc8b9\") pod \"cinder-db-create-257lm\" (UID: \"fdb459dd-26c4-49c4-bf1f-dad807314e3d\") " pod="openstack/cinder-db-create-257lm" Oct 11 04:07:12 crc kubenswrapper[4967]: I1011 04:07:12.876357 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-f5bhd" Oct 11 04:07:12 crc kubenswrapper[4967]: I1011 04:07:12.916765 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-257lm" Oct 11 04:07:12 crc kubenswrapper[4967]: I1011 04:07:12.945004 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xsjk5\" (UniqueName: \"kubernetes.io/projected/d9c71007-45ce-4684-87eb-1884ec5192c9-kube-api-access-xsjk5\") pod \"neutron-db-create-9p6rn\" (UID: \"d9c71007-45ce-4684-87eb-1884ec5192c9\") " pod="openstack/neutron-db-create-9p6rn" Oct 11 04:07:12 crc kubenswrapper[4967]: I1011 04:07:12.952430 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-db-sync-ks59l"] Oct 11 04:07:12 crc kubenswrapper[4967]: I1011 04:07:12.953709 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ks59l" Oct 11 04:07:12 crc kubenswrapper[4967]: I1011 04:07:12.958742 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 11 04:07:12 crc kubenswrapper[4967]: I1011 04:07:12.958973 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-k7kxn" Oct 11 04:07:12 crc kubenswrapper[4967]: I1011 04:07:12.960307 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 11 04:07:12 crc kubenswrapper[4967]: I1011 04:07:12.963344 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 11 04:07:12 crc kubenswrapper[4967]: I1011 04:07:12.966834 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-ks59l"] Oct 11 04:07:13 crc kubenswrapper[4967]: I1011 04:07:13.046888 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ccef7a0-eebe-4794-b38d-690a98ee2055-combined-ca-bundle\") pod \"keystone-db-sync-ks59l\" (UID: \"0ccef7a0-eebe-4794-b38d-690a98ee2055\") " pod="openstack/keystone-db-sync-ks59l" Oct 11 04:07:13 crc kubenswrapper[4967]: I1011 04:07:13.046962 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xsjk5\" (UniqueName: \"kubernetes.io/projected/d9c71007-45ce-4684-87eb-1884ec5192c9-kube-api-access-xsjk5\") pod \"neutron-db-create-9p6rn\" (UID: \"d9c71007-45ce-4684-87eb-1884ec5192c9\") " pod="openstack/neutron-db-create-9p6rn" Oct 11 04:07:13 crc kubenswrapper[4967]: I1011 04:07:13.047020 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ccef7a0-eebe-4794-b38d-690a98ee2055-config-data\") pod \"keystone-db-sync-ks59l\" (UID: \"0ccef7a0-eebe-4794-b38d-690a98ee2055\") " pod="openstack/keystone-db-sync-ks59l" Oct 11 04:07:13 crc kubenswrapper[4967]: I1011 04:07:13.047149 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hnz5m\" (UniqueName: \"kubernetes.io/projected/0ccef7a0-eebe-4794-b38d-690a98ee2055-kube-api-access-hnz5m\") pod \"keystone-db-sync-ks59l\" (UID: \"0ccef7a0-eebe-4794-b38d-690a98ee2055\") " pod="openstack/keystone-db-sync-ks59l" Oct 11 04:07:13 crc kubenswrapper[4967]: I1011 04:07:13.065482 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xsjk5\" (UniqueName: \"kubernetes.io/projected/d9c71007-45ce-4684-87eb-1884ec5192c9-kube-api-access-xsjk5\") pod \"neutron-db-create-9p6rn\" (UID: \"d9c71007-45ce-4684-87eb-1884ec5192c9\") " pod="openstack/neutron-db-create-9p6rn" Oct 11 04:07:13 crc kubenswrapper[4967]: I1011 04:07:13.141947 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9p6rn" Oct 11 04:07:13 crc kubenswrapper[4967]: I1011 04:07:13.148592 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ccef7a0-eebe-4794-b38d-690a98ee2055-combined-ca-bundle\") pod \"keystone-db-sync-ks59l\" (UID: \"0ccef7a0-eebe-4794-b38d-690a98ee2055\") " pod="openstack/keystone-db-sync-ks59l" Oct 11 04:07:13 crc kubenswrapper[4967]: I1011 04:07:13.148664 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ccef7a0-eebe-4794-b38d-690a98ee2055-config-data\") pod \"keystone-db-sync-ks59l\" (UID: \"0ccef7a0-eebe-4794-b38d-690a98ee2055\") " pod="openstack/keystone-db-sync-ks59l" Oct 11 04:07:13 crc kubenswrapper[4967]: I1011 04:07:13.148737 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hnz5m\" (UniqueName: \"kubernetes.io/projected/0ccef7a0-eebe-4794-b38d-690a98ee2055-kube-api-access-hnz5m\") pod \"keystone-db-sync-ks59l\" (UID: \"0ccef7a0-eebe-4794-b38d-690a98ee2055\") " pod="openstack/keystone-db-sync-ks59l" Oct 11 04:07:13 crc kubenswrapper[4967]: I1011 04:07:13.152223 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ccef7a0-eebe-4794-b38d-690a98ee2055-combined-ca-bundle\") pod \"keystone-db-sync-ks59l\" (UID: \"0ccef7a0-eebe-4794-b38d-690a98ee2055\") " pod="openstack/keystone-db-sync-ks59l" Oct 11 04:07:13 crc kubenswrapper[4967]: I1011 04:07:13.162767 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ccef7a0-eebe-4794-b38d-690a98ee2055-config-data\") pod \"keystone-db-sync-ks59l\" (UID: \"0ccef7a0-eebe-4794-b38d-690a98ee2055\") " pod="openstack/keystone-db-sync-ks59l" Oct 11 04:07:13 crc kubenswrapper[4967]: I1011 04:07:13.176604 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hnz5m\" (UniqueName: \"kubernetes.io/projected/0ccef7a0-eebe-4794-b38d-690a98ee2055-kube-api-access-hnz5m\") pod \"keystone-db-sync-ks59l\" (UID: \"0ccef7a0-eebe-4794-b38d-690a98ee2055\") " pod="openstack/keystone-db-sync-ks59l" Oct 11 04:07:13 crc kubenswrapper[4967]: I1011 04:07:13.276478 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ks59l" Oct 11 04:07:17 crc kubenswrapper[4967]: I1011 04:07:17.246792 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-create-f5bhd"] Oct 11 04:07:17 crc kubenswrapper[4967]: W1011 04:07:17.254234 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod6f16161f_b863_46ca_b2f1_7a6248f27c4d.slice/crio-3744d6a27ccccd36c399a3d7b4686809c84149f8f337ae24919d1fbec8fe18f4 WatchSource:0}: Error finding container 3744d6a27ccccd36c399a3d7b4686809c84149f8f337ae24919d1fbec8fe18f4: Status 404 returned error can't find the container with id 3744d6a27ccccd36c399a3d7b4686809c84149f8f337ae24919d1fbec8fe18f4 Oct 11 04:07:17 crc kubenswrapper[4967]: I1011 04:07:17.343585 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-create-257lm"] Oct 11 04:07:17 crc kubenswrapper[4967]: W1011 04:07:17.348857 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfdb459dd_26c4_49c4_bf1f_dad807314e3d.slice/crio-511cc2b809ca767a3d51a27e85363caa26ca11e45ef3f7c548d9cf3210398424 WatchSource:0}: Error finding container 511cc2b809ca767a3d51a27e85363caa26ca11e45ef3f7c548d9cf3210398424: Status 404 returned error can't find the container with id 511cc2b809ca767a3d51a27e85363caa26ca11e45ef3f7c548d9cf3210398424 Oct 11 04:07:17 crc kubenswrapper[4967]: I1011 04:07:17.404719 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-db-sync-ks59l"] Oct 11 04:07:17 crc kubenswrapper[4967]: W1011 04:07:17.421243 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod0ccef7a0_eebe_4794_b38d_690a98ee2055.slice/crio-135419cd51312a93f574965953429b3364518e34b03dabcb160e9267c8a3877b WatchSource:0}: Error finding container 135419cd51312a93f574965953429b3364518e34b03dabcb160e9267c8a3877b: Status 404 returned error can't find the container with id 135419cd51312a93f574965953429b3364518e34b03dabcb160e9267c8a3877b Oct 11 04:07:17 crc kubenswrapper[4967]: I1011 04:07:17.479774 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-create-9p6rn"] Oct 11 04:07:17 crc kubenswrapper[4967]: W1011 04:07:17.496390 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podd9c71007_45ce_4684_87eb_1884ec5192c9.slice/crio-d46d0fc7a54ed6be23c8eb89440c254271b00647bb2dc0729a856c6f0e60dcaf WatchSource:0}: Error finding container d46d0fc7a54ed6be23c8eb89440c254271b00647bb2dc0729a856c6f0e60dcaf: Status 404 returned error can't find the container with id d46d0fc7a54ed6be23c8eb89440c254271b00647bb2dc0729a856c6f0e60dcaf Oct 11 04:07:17 crc kubenswrapper[4967]: I1011 04:07:17.565030 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-storage-0"] Oct 11 04:07:17 crc kubenswrapper[4967]: I1011 04:07:17.745831 4967 generic.go:334] "Generic (PLEG): container finished" podID="6f16161f-b863-46ca-b2f1-7a6248f27c4d" containerID="99d6af2f22ee5cb47dba8ad27385fb8bd4ee168973097919bb99a148169b6a28" exitCode=0 Oct 11 04:07:17 crc kubenswrapper[4967]: I1011 04:07:17.745941 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-f5bhd" event={"ID":"6f16161f-b863-46ca-b2f1-7a6248f27c4d","Type":"ContainerDied","Data":"99d6af2f22ee5cb47dba8ad27385fb8bd4ee168973097919bb99a148169b6a28"} Oct 11 04:07:17 crc kubenswrapper[4967]: I1011 04:07:17.745992 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-f5bhd" event={"ID":"6f16161f-b863-46ca-b2f1-7a6248f27c4d","Type":"ContainerStarted","Data":"3744d6a27ccccd36c399a3d7b4686809c84149f8f337ae24919d1fbec8fe18f4"} Oct 11 04:07:17 crc kubenswrapper[4967]: I1011 04:07:17.747552 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ks59l" event={"ID":"0ccef7a0-eebe-4794-b38d-690a98ee2055","Type":"ContainerStarted","Data":"135419cd51312a93f574965953429b3364518e34b03dabcb160e9267c8a3877b"} Oct 11 04:07:17 crc kubenswrapper[4967]: I1011 04:07:17.749750 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9p6rn" event={"ID":"d9c71007-45ce-4684-87eb-1884ec5192c9","Type":"ContainerStarted","Data":"d63c539df8be1bba90129b4002c49f274647e7342820081417c3303fbbfaead7"} Oct 11 04:07:17 crc kubenswrapper[4967]: I1011 04:07:17.749907 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9p6rn" event={"ID":"d9c71007-45ce-4684-87eb-1884ec5192c9","Type":"ContainerStarted","Data":"d46d0fc7a54ed6be23c8eb89440c254271b00647bb2dc0729a856c6f0e60dcaf"} Oct 11 04:07:17 crc kubenswrapper[4967]: I1011 04:07:17.751622 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-257lm" event={"ID":"fdb459dd-26c4-49c4-bf1f-dad807314e3d","Type":"ContainerStarted","Data":"e76735772140c83529136bb7bb0ee17f72e4d64e36fa332ded89a7ba6254197b"} Oct 11 04:07:17 crc kubenswrapper[4967]: I1011 04:07:17.751799 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-257lm" event={"ID":"fdb459dd-26c4-49c4-bf1f-dad807314e3d","Type":"ContainerStarted","Data":"511cc2b809ca767a3d51a27e85363caa26ca11e45ef3f7c548d9cf3210398424"} Oct 11 04:07:17 crc kubenswrapper[4967]: I1011 04:07:17.757747 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c09e3333-ad4d-495f-8973-7e2d2f0931d4","Type":"ContainerStarted","Data":"f4ea2380f773be0bd6615a93f2e20067696f4910f5726c486348c21baacc54f4"} Oct 11 04:07:17 crc kubenswrapper[4967]: I1011 04:07:17.760177 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-mgcb9" event={"ID":"df08696a-e0e9-4d45-9ffb-c41eff1b5fb4","Type":"ContainerStarted","Data":"dc76d506020f9eaa727c4fa0b31d77be06538ced515178c9fdefb0d2a99bb700"} Oct 11 04:07:17 crc kubenswrapper[4967]: I1011 04:07:17.790990 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-create-257lm" podStartSLOduration=5.790969848 podStartE2EDuration="5.790969848s" podCreationTimestamp="2025-10-11 04:07:12 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:07:17.784543686 +0000 UTC m=+965.747752619" watchObservedRunningTime="2025-10-11 04:07:17.790969848 +0000 UTC m=+965.754178781" Oct 11 04:07:17 crc kubenswrapper[4967]: I1011 04:07:17.806250 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-db-sync-mgcb9" podStartSLOduration=1.94441977 podStartE2EDuration="12.806234908s" podCreationTimestamp="2025-10-11 04:07:05 +0000 UTC" firstStartedPulling="2025-10-11 04:07:06.110588441 +0000 UTC m=+954.073797374" lastFinishedPulling="2025-10-11 04:07:16.972403549 +0000 UTC m=+964.935612512" observedRunningTime="2025-10-11 04:07:17.802555411 +0000 UTC m=+965.765764364" watchObservedRunningTime="2025-10-11 04:07:17.806234908 +0000 UTC m=+965.769443841" Oct 11 04:07:18 crc kubenswrapper[4967]: I1011 04:07:18.768640 4967 generic.go:334] "Generic (PLEG): container finished" podID="fdb459dd-26c4-49c4-bf1f-dad807314e3d" containerID="e76735772140c83529136bb7bb0ee17f72e4d64e36fa332ded89a7ba6254197b" exitCode=0 Oct 11 04:07:18 crc kubenswrapper[4967]: I1011 04:07:18.768693 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-257lm" event={"ID":"fdb459dd-26c4-49c4-bf1f-dad807314e3d","Type":"ContainerDied","Data":"e76735772140c83529136bb7bb0ee17f72e4d64e36fa332ded89a7ba6254197b"} Oct 11 04:07:18 crc kubenswrapper[4967]: I1011 04:07:18.770927 4967 generic.go:334] "Generic (PLEG): container finished" podID="d9c71007-45ce-4684-87eb-1884ec5192c9" containerID="d63c539df8be1bba90129b4002c49f274647e7342820081417c3303fbbfaead7" exitCode=0 Oct 11 04:07:18 crc kubenswrapper[4967]: I1011 04:07:18.770958 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9p6rn" event={"ID":"d9c71007-45ce-4684-87eb-1884ec5192c9","Type":"ContainerDied","Data":"d63c539df8be1bba90129b4002c49f274647e7342820081417c3303fbbfaead7"} Oct 11 04:07:19 crc kubenswrapper[4967]: I1011 04:07:19.783521 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-create-f5bhd" event={"ID":"6f16161f-b863-46ca-b2f1-7a6248f27c4d","Type":"ContainerDied","Data":"3744d6a27ccccd36c399a3d7b4686809c84149f8f337ae24919d1fbec8fe18f4"} Oct 11 04:07:19 crc kubenswrapper[4967]: I1011 04:07:19.783823 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3744d6a27ccccd36c399a3d7b4686809c84149f8f337ae24919d1fbec8fe18f4" Oct 11 04:07:19 crc kubenswrapper[4967]: I1011 04:07:19.829172 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-f5bhd" Oct 11 04:07:19 crc kubenswrapper[4967]: I1011 04:07:19.963899 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r86rl\" (UniqueName: \"kubernetes.io/projected/6f16161f-b863-46ca-b2f1-7a6248f27c4d-kube-api-access-r86rl\") pod \"6f16161f-b863-46ca-b2f1-7a6248f27c4d\" (UID: \"6f16161f-b863-46ca-b2f1-7a6248f27c4d\") " Oct 11 04:07:19 crc kubenswrapper[4967]: I1011 04:07:19.978414 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6f16161f-b863-46ca-b2f1-7a6248f27c4d-kube-api-access-r86rl" (OuterVolumeSpecName: "kube-api-access-r86rl") pod "6f16161f-b863-46ca-b2f1-7a6248f27c4d" (UID: "6f16161f-b863-46ca-b2f1-7a6248f27c4d"). InnerVolumeSpecName "kube-api-access-r86rl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:07:20 crc kubenswrapper[4967]: I1011 04:07:20.066474 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r86rl\" (UniqueName: \"kubernetes.io/projected/6f16161f-b863-46ca-b2f1-7a6248f27c4d-kube-api-access-r86rl\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:20 crc kubenswrapper[4967]: I1011 04:07:20.791222 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-create-f5bhd" Oct 11 04:07:22 crc kubenswrapper[4967]: I1011 04:07:22.422446 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-257lm" Oct 11 04:07:22 crc kubenswrapper[4967]: I1011 04:07:22.452817 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9p6rn" Oct 11 04:07:22 crc kubenswrapper[4967]: I1011 04:07:22.503923 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tc8b9\" (UniqueName: \"kubernetes.io/projected/fdb459dd-26c4-49c4-bf1f-dad807314e3d-kube-api-access-tc8b9\") pod \"fdb459dd-26c4-49c4-bf1f-dad807314e3d\" (UID: \"fdb459dd-26c4-49c4-bf1f-dad807314e3d\") " Oct 11 04:07:22 crc kubenswrapper[4967]: I1011 04:07:22.503991 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xsjk5\" (UniqueName: \"kubernetes.io/projected/d9c71007-45ce-4684-87eb-1884ec5192c9-kube-api-access-xsjk5\") pod \"d9c71007-45ce-4684-87eb-1884ec5192c9\" (UID: \"d9c71007-45ce-4684-87eb-1884ec5192c9\") " Oct 11 04:07:22 crc kubenswrapper[4967]: I1011 04:07:22.507886 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fdb459dd-26c4-49c4-bf1f-dad807314e3d-kube-api-access-tc8b9" (OuterVolumeSpecName: "kube-api-access-tc8b9") pod "fdb459dd-26c4-49c4-bf1f-dad807314e3d" (UID: "fdb459dd-26c4-49c4-bf1f-dad807314e3d"). InnerVolumeSpecName "kube-api-access-tc8b9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:07:22 crc kubenswrapper[4967]: I1011 04:07:22.508463 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d9c71007-45ce-4684-87eb-1884ec5192c9-kube-api-access-xsjk5" (OuterVolumeSpecName: "kube-api-access-xsjk5") pod "d9c71007-45ce-4684-87eb-1884ec5192c9" (UID: "d9c71007-45ce-4684-87eb-1884ec5192c9"). InnerVolumeSpecName "kube-api-access-xsjk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:07:22 crc kubenswrapper[4967]: I1011 04:07:22.605216 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tc8b9\" (UniqueName: \"kubernetes.io/projected/fdb459dd-26c4-49c4-bf1f-dad807314e3d-kube-api-access-tc8b9\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:22 crc kubenswrapper[4967]: I1011 04:07:22.605261 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xsjk5\" (UniqueName: \"kubernetes.io/projected/d9c71007-45ce-4684-87eb-1884ec5192c9-kube-api-access-xsjk5\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:22 crc kubenswrapper[4967]: I1011 04:07:22.810020 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ks59l" event={"ID":"0ccef7a0-eebe-4794-b38d-690a98ee2055","Type":"ContainerStarted","Data":"05e2a28172819cddfb09e990e95790043afc1033045364b6fd724ba8b52a2b0b"} Oct 11 04:07:22 crc kubenswrapper[4967]: I1011 04:07:22.823083 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-create-9p6rn" Oct 11 04:07:22 crc kubenswrapper[4967]: I1011 04:07:22.824866 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-create-257lm" Oct 11 04:07:22 crc kubenswrapper[4967]: I1011 04:07:22.834470 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-db-sync-ks59l" podStartSLOduration=5.973398835 podStartE2EDuration="10.834445474s" podCreationTimestamp="2025-10-11 04:07:12 +0000 UTC" firstStartedPulling="2025-10-11 04:07:17.423486061 +0000 UTC m=+965.386694994" lastFinishedPulling="2025-10-11 04:07:22.28453269 +0000 UTC m=+970.247741633" observedRunningTime="2025-10-11 04:07:22.828581716 +0000 UTC m=+970.791790649" watchObservedRunningTime="2025-10-11 04:07:22.834445474 +0000 UTC m=+970.797654407" Oct 11 04:07:22 crc kubenswrapper[4967]: I1011 04:07:22.841961 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-create-9p6rn" event={"ID":"d9c71007-45ce-4684-87eb-1884ec5192c9","Type":"ContainerDied","Data":"d46d0fc7a54ed6be23c8eb89440c254271b00647bb2dc0729a856c6f0e60dcaf"} Oct 11 04:07:22 crc kubenswrapper[4967]: I1011 04:07:22.841998 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d46d0fc7a54ed6be23c8eb89440c254271b00647bb2dc0729a856c6f0e60dcaf" Oct 11 04:07:22 crc kubenswrapper[4967]: I1011 04:07:22.842009 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-create-257lm" event={"ID":"fdb459dd-26c4-49c4-bf1f-dad807314e3d","Type":"ContainerDied","Data":"511cc2b809ca767a3d51a27e85363caa26ca11e45ef3f7c548d9cf3210398424"} Oct 11 04:07:22 crc kubenswrapper[4967]: I1011 04:07:22.842018 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="511cc2b809ca767a3d51a27e85363caa26ca11e45ef3f7c548d9cf3210398424" Oct 11 04:07:22 crc kubenswrapper[4967]: I1011 04:07:22.842025 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c09e3333-ad4d-495f-8973-7e2d2f0931d4","Type":"ContainerStarted","Data":"27e16f3f25e78dc9077f02e6a71f480964105d49a4b2560af892b17e29d6e812"} Oct 11 04:07:22 crc kubenswrapper[4967]: I1011 04:07:22.842034 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c09e3333-ad4d-495f-8973-7e2d2f0931d4","Type":"ContainerStarted","Data":"da5555a365f90714053e8c6960a81c40767f381b6971b62a2293fd7df5386120"} Oct 11 04:07:23 crc kubenswrapper[4967]: I1011 04:07:23.848930 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c09e3333-ad4d-495f-8973-7e2d2f0931d4","Type":"ContainerStarted","Data":"ff7e1b871c089c8ab43c659282ba0e83fe1f835fc518a2258b1d91138836fdb2"} Oct 11 04:07:23 crc kubenswrapper[4967]: I1011 04:07:23.856273 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c09e3333-ad4d-495f-8973-7e2d2f0931d4","Type":"ContainerStarted","Data":"0ec2ade05cf307d772572d8d8f7947da2ace1b95f1e209e62f5f4b16381df908"} Oct 11 04:07:24 crc kubenswrapper[4967]: I1011 04:07:24.858579 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c09e3333-ad4d-495f-8973-7e2d2f0931d4","Type":"ContainerStarted","Data":"33c91482bda04b1f447ba086b9400edce88ac0f19fdd6db645a83ff4f22ebc7a"} Oct 11 04:07:24 crc kubenswrapper[4967]: I1011 04:07:24.858922 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c09e3333-ad4d-495f-8973-7e2d2f0931d4","Type":"ContainerStarted","Data":"674f9f121c406a01ae6ffe79d0c6f99be0eca0f1763ee408d1e1a9afb558bbba"} Oct 11 04:07:24 crc kubenswrapper[4967]: I1011 04:07:24.860249 4967 generic.go:334] "Generic (PLEG): container finished" podID="df08696a-e0e9-4d45-9ffb-c41eff1b5fb4" containerID="dc76d506020f9eaa727c4fa0b31d77be06538ced515178c9fdefb0d2a99bb700" exitCode=0 Oct 11 04:07:24 crc kubenswrapper[4967]: I1011 04:07:24.860271 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-mgcb9" event={"ID":"df08696a-e0e9-4d45-9ffb-c41eff1b5fb4","Type":"ContainerDied","Data":"dc76d506020f9eaa727c4fa0b31d77be06538ced515178c9fdefb0d2a99bb700"} Oct 11 04:07:25 crc kubenswrapper[4967]: I1011 04:07:25.886883 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c09e3333-ad4d-495f-8973-7e2d2f0931d4","Type":"ContainerStarted","Data":"76c0be0fa0c1ac09155795b78a862ec55e1e1635f4ad84208c0390016c87d62f"} Oct 11 04:07:25 crc kubenswrapper[4967]: I1011 04:07:25.887159 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c09e3333-ad4d-495f-8973-7e2d2f0931d4","Type":"ContainerStarted","Data":"bc9145aeed5204ca532f40f6dead9bf49f9e87b9c46f58b1898f0ff92316ad9b"} Oct 11 04:07:25 crc kubenswrapper[4967]: I1011 04:07:25.892667 4967 generic.go:334] "Generic (PLEG): container finished" podID="0ccef7a0-eebe-4794-b38d-690a98ee2055" containerID="05e2a28172819cddfb09e990e95790043afc1033045364b6fd724ba8b52a2b0b" exitCode=0 Oct 11 04:07:25 crc kubenswrapper[4967]: I1011 04:07:25.892987 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ks59l" event={"ID":"0ccef7a0-eebe-4794-b38d-690a98ee2055","Type":"ContainerDied","Data":"05e2a28172819cddfb09e990e95790043afc1033045364b6fd724ba8b52a2b0b"} Oct 11 04:07:26 crc kubenswrapper[4967]: I1011 04:07:26.391197 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-mgcb9" Oct 11 04:07:26 crc kubenswrapper[4967]: I1011 04:07:26.573323 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df08696a-e0e9-4d45-9ffb-c41eff1b5fb4-config-data\") pod \"df08696a-e0e9-4d45-9ffb-c41eff1b5fb4\" (UID: \"df08696a-e0e9-4d45-9ffb-c41eff1b5fb4\") " Oct 11 04:07:26 crc kubenswrapper[4967]: I1011 04:07:26.574206 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/df08696a-e0e9-4d45-9ffb-c41eff1b5fb4-db-sync-config-data\") pod \"df08696a-e0e9-4d45-9ffb-c41eff1b5fb4\" (UID: \"df08696a-e0e9-4d45-9ffb-c41eff1b5fb4\") " Oct 11 04:07:26 crc kubenswrapper[4967]: I1011 04:07:26.574696 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkbwn\" (UniqueName: \"kubernetes.io/projected/df08696a-e0e9-4d45-9ffb-c41eff1b5fb4-kube-api-access-jkbwn\") pod \"df08696a-e0e9-4d45-9ffb-c41eff1b5fb4\" (UID: \"df08696a-e0e9-4d45-9ffb-c41eff1b5fb4\") " Oct 11 04:07:26 crc kubenswrapper[4967]: I1011 04:07:26.574859 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df08696a-e0e9-4d45-9ffb-c41eff1b5fb4-combined-ca-bundle\") pod \"df08696a-e0e9-4d45-9ffb-c41eff1b5fb4\" (UID: \"df08696a-e0e9-4d45-9ffb-c41eff1b5fb4\") " Oct 11 04:07:26 crc kubenswrapper[4967]: I1011 04:07:26.578169 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/df08696a-e0e9-4d45-9ffb-c41eff1b5fb4-kube-api-access-jkbwn" (OuterVolumeSpecName: "kube-api-access-jkbwn") pod "df08696a-e0e9-4d45-9ffb-c41eff1b5fb4" (UID: "df08696a-e0e9-4d45-9ffb-c41eff1b5fb4"). InnerVolumeSpecName "kube-api-access-jkbwn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:07:26 crc kubenswrapper[4967]: I1011 04:07:26.578908 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df08696a-e0e9-4d45-9ffb-c41eff1b5fb4-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "df08696a-e0e9-4d45-9ffb-c41eff1b5fb4" (UID: "df08696a-e0e9-4d45-9ffb-c41eff1b5fb4"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:07:26 crc kubenswrapper[4967]: I1011 04:07:26.597025 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df08696a-e0e9-4d45-9ffb-c41eff1b5fb4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "df08696a-e0e9-4d45-9ffb-c41eff1b5fb4" (UID: "df08696a-e0e9-4d45-9ffb-c41eff1b5fb4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:07:26 crc kubenswrapper[4967]: I1011 04:07:26.619385 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/df08696a-e0e9-4d45-9ffb-c41eff1b5fb4-config-data" (OuterVolumeSpecName: "config-data") pod "df08696a-e0e9-4d45-9ffb-c41eff1b5fb4" (UID: "df08696a-e0e9-4d45-9ffb-c41eff1b5fb4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:07:26 crc kubenswrapper[4967]: I1011 04:07:26.677612 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/df08696a-e0e9-4d45-9ffb-c41eff1b5fb4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:26 crc kubenswrapper[4967]: I1011 04:07:26.677667 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/df08696a-e0e9-4d45-9ffb-c41eff1b5fb4-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:26 crc kubenswrapper[4967]: I1011 04:07:26.677680 4967 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/df08696a-e0e9-4d45-9ffb-c41eff1b5fb4-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:26 crc kubenswrapper[4967]: I1011 04:07:26.677693 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkbwn\" (UniqueName: \"kubernetes.io/projected/df08696a-e0e9-4d45-9ffb-c41eff1b5fb4-kube-api-access-jkbwn\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:26 crc kubenswrapper[4967]: I1011 04:07:26.917371 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-db-sync-mgcb9" event={"ID":"df08696a-e0e9-4d45-9ffb-c41eff1b5fb4","Type":"ContainerDied","Data":"9f149671ac7f0a791e3adb6f8dd3664bd27ad1cc13b8f0e843410cb1225c5201"} Oct 11 04:07:26 crc kubenswrapper[4967]: I1011 04:07:26.917402 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9f149671ac7f0a791e3adb6f8dd3664bd27ad1cc13b8f0e843410cb1225c5201" Oct 11 04:07:26 crc kubenswrapper[4967]: I1011 04:07:26.917405 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-db-sync-mgcb9" Oct 11 04:07:26 crc kubenswrapper[4967]: I1011 04:07:26.928653 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c09e3333-ad4d-495f-8973-7e2d2f0931d4","Type":"ContainerStarted","Data":"06409f56a8a064fed803f0a6b7653cb771e60451faf333780a7def96bec4c9cc"} Oct 11 04:07:26 crc kubenswrapper[4967]: I1011 04:07:26.928702 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c09e3333-ad4d-495f-8973-7e2d2f0931d4","Type":"ContainerStarted","Data":"34d279062f2c5a7c4c872dd024c5f81fcbdda1087bb82750d07c07953da3e87e"} Oct 11 04:07:26 crc kubenswrapper[4967]: I1011 04:07:26.928715 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c09e3333-ad4d-495f-8973-7e2d2f0931d4","Type":"ContainerStarted","Data":"9e9282b3abbc90b1e23003da5adef902fe2bfcd42f6782fe2a36534b92bf1cac"} Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.209334 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ks59l" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.306597 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-sp4wt"] Oct 11 04:07:27 crc kubenswrapper[4967]: E1011 04:07:27.306899 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="df08696a-e0e9-4d45-9ffb-c41eff1b5fb4" containerName="glance-db-sync" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.306915 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="df08696a-e0e9-4d45-9ffb-c41eff1b5fb4" containerName="glance-db-sync" Oct 11 04:07:27 crc kubenswrapper[4967]: E1011 04:07:27.306931 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="fdb459dd-26c4-49c4-bf1f-dad807314e3d" containerName="mariadb-database-create" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.306939 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="fdb459dd-26c4-49c4-bf1f-dad807314e3d" containerName="mariadb-database-create" Oct 11 04:07:27 crc kubenswrapper[4967]: E1011 04:07:27.306949 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d9c71007-45ce-4684-87eb-1884ec5192c9" containerName="mariadb-database-create" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.306965 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="d9c71007-45ce-4684-87eb-1884ec5192c9" containerName="mariadb-database-create" Oct 11 04:07:27 crc kubenswrapper[4967]: E1011 04:07:27.306976 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6f16161f-b863-46ca-b2f1-7a6248f27c4d" containerName="mariadb-database-create" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.306982 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="6f16161f-b863-46ca-b2f1-7a6248f27c4d" containerName="mariadb-database-create" Oct 11 04:07:27 crc kubenswrapper[4967]: E1011 04:07:27.306992 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0ccef7a0-eebe-4794-b38d-690a98ee2055" containerName="keystone-db-sync" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.306997 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="0ccef7a0-eebe-4794-b38d-690a98ee2055" containerName="keystone-db-sync" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.307171 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="fdb459dd-26c4-49c4-bf1f-dad807314e3d" containerName="mariadb-database-create" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.307190 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="0ccef7a0-eebe-4794-b38d-690a98ee2055" containerName="keystone-db-sync" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.307214 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="d9c71007-45ce-4684-87eb-1884ec5192c9" containerName="mariadb-database-create" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.307225 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="6f16161f-b863-46ca-b2f1-7a6248f27c4d" containerName="mariadb-database-create" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.307233 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="df08696a-e0e9-4d45-9ffb-c41eff1b5fb4" containerName="glance-db-sync" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.307990 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74dc88fc-sp4wt" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.326778 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-sp4wt"] Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.405574 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hnz5m\" (UniqueName: \"kubernetes.io/projected/0ccef7a0-eebe-4794-b38d-690a98ee2055-kube-api-access-hnz5m\") pod \"0ccef7a0-eebe-4794-b38d-690a98ee2055\" (UID: \"0ccef7a0-eebe-4794-b38d-690a98ee2055\") " Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.405611 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ccef7a0-eebe-4794-b38d-690a98ee2055-combined-ca-bundle\") pod \"0ccef7a0-eebe-4794-b38d-690a98ee2055\" (UID: \"0ccef7a0-eebe-4794-b38d-690a98ee2055\") " Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.405658 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ccef7a0-eebe-4794-b38d-690a98ee2055-config-data\") pod \"0ccef7a0-eebe-4794-b38d-690a98ee2055\" (UID: \"0ccef7a0-eebe-4794-b38d-690a98ee2055\") " Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.405989 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kljb7\" (UniqueName: \"kubernetes.io/projected/43f1cabd-1a2d-4996-98e4-d90f58007fab-kube-api-access-kljb7\") pod \"dnsmasq-dns-74dc88fc-sp4wt\" (UID: \"43f1cabd-1a2d-4996-98e4-d90f58007fab\") " pod="openstack/dnsmasq-dns-74dc88fc-sp4wt" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.406021 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/43f1cabd-1a2d-4996-98e4-d90f58007fab-ovsdbserver-nb\") pod \"dnsmasq-dns-74dc88fc-sp4wt\" (UID: \"43f1cabd-1a2d-4996-98e4-d90f58007fab\") " pod="openstack/dnsmasq-dns-74dc88fc-sp4wt" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.406091 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43f1cabd-1a2d-4996-98e4-d90f58007fab-config\") pod \"dnsmasq-dns-74dc88fc-sp4wt\" (UID: \"43f1cabd-1a2d-4996-98e4-d90f58007fab\") " pod="openstack/dnsmasq-dns-74dc88fc-sp4wt" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.406117 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43f1cabd-1a2d-4996-98e4-d90f58007fab-dns-svc\") pod \"dnsmasq-dns-74dc88fc-sp4wt\" (UID: \"43f1cabd-1a2d-4996-98e4-d90f58007fab\") " pod="openstack/dnsmasq-dns-74dc88fc-sp4wt" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.406160 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/43f1cabd-1a2d-4996-98e4-d90f58007fab-ovsdbserver-sb\") pod \"dnsmasq-dns-74dc88fc-sp4wt\" (UID: \"43f1cabd-1a2d-4996-98e4-d90f58007fab\") " pod="openstack/dnsmasq-dns-74dc88fc-sp4wt" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.424094 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0ccef7a0-eebe-4794-b38d-690a98ee2055-kube-api-access-hnz5m" (OuterVolumeSpecName: "kube-api-access-hnz5m") pod "0ccef7a0-eebe-4794-b38d-690a98ee2055" (UID: "0ccef7a0-eebe-4794-b38d-690a98ee2055"). InnerVolumeSpecName "kube-api-access-hnz5m". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.443929 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ccef7a0-eebe-4794-b38d-690a98ee2055-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "0ccef7a0-eebe-4794-b38d-690a98ee2055" (UID: "0ccef7a0-eebe-4794-b38d-690a98ee2055"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.494316 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0ccef7a0-eebe-4794-b38d-690a98ee2055-config-data" (OuterVolumeSpecName: "config-data") pod "0ccef7a0-eebe-4794-b38d-690a98ee2055" (UID: "0ccef7a0-eebe-4794-b38d-690a98ee2055"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.507699 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43f1cabd-1a2d-4996-98e4-d90f58007fab-config\") pod \"dnsmasq-dns-74dc88fc-sp4wt\" (UID: \"43f1cabd-1a2d-4996-98e4-d90f58007fab\") " pod="openstack/dnsmasq-dns-74dc88fc-sp4wt" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.507751 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43f1cabd-1a2d-4996-98e4-d90f58007fab-dns-svc\") pod \"dnsmasq-dns-74dc88fc-sp4wt\" (UID: \"43f1cabd-1a2d-4996-98e4-d90f58007fab\") " pod="openstack/dnsmasq-dns-74dc88fc-sp4wt" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.507799 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/43f1cabd-1a2d-4996-98e4-d90f58007fab-ovsdbserver-sb\") pod \"dnsmasq-dns-74dc88fc-sp4wt\" (UID: \"43f1cabd-1a2d-4996-98e4-d90f58007fab\") " pod="openstack/dnsmasq-dns-74dc88fc-sp4wt" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.507873 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kljb7\" (UniqueName: \"kubernetes.io/projected/43f1cabd-1a2d-4996-98e4-d90f58007fab-kube-api-access-kljb7\") pod \"dnsmasq-dns-74dc88fc-sp4wt\" (UID: \"43f1cabd-1a2d-4996-98e4-d90f58007fab\") " pod="openstack/dnsmasq-dns-74dc88fc-sp4wt" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.507895 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/43f1cabd-1a2d-4996-98e4-d90f58007fab-ovsdbserver-nb\") pod \"dnsmasq-dns-74dc88fc-sp4wt\" (UID: \"43f1cabd-1a2d-4996-98e4-d90f58007fab\") " pod="openstack/dnsmasq-dns-74dc88fc-sp4wt" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.507949 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hnz5m\" (UniqueName: \"kubernetes.io/projected/0ccef7a0-eebe-4794-b38d-690a98ee2055-kube-api-access-hnz5m\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.507966 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/0ccef7a0-eebe-4794-b38d-690a98ee2055-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.507975 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/0ccef7a0-eebe-4794-b38d-690a98ee2055-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.508833 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43f1cabd-1a2d-4996-98e4-d90f58007fab-dns-svc\") pod \"dnsmasq-dns-74dc88fc-sp4wt\" (UID: \"43f1cabd-1a2d-4996-98e4-d90f58007fab\") " pod="openstack/dnsmasq-dns-74dc88fc-sp4wt" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.508842 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43f1cabd-1a2d-4996-98e4-d90f58007fab-config\") pod \"dnsmasq-dns-74dc88fc-sp4wt\" (UID: \"43f1cabd-1a2d-4996-98e4-d90f58007fab\") " pod="openstack/dnsmasq-dns-74dc88fc-sp4wt" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.508886 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/43f1cabd-1a2d-4996-98e4-d90f58007fab-ovsdbserver-nb\") pod \"dnsmasq-dns-74dc88fc-sp4wt\" (UID: \"43f1cabd-1a2d-4996-98e4-d90f58007fab\") " pod="openstack/dnsmasq-dns-74dc88fc-sp4wt" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.509363 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/43f1cabd-1a2d-4996-98e4-d90f58007fab-ovsdbserver-sb\") pod \"dnsmasq-dns-74dc88fc-sp4wt\" (UID: \"43f1cabd-1a2d-4996-98e4-d90f58007fab\") " pod="openstack/dnsmasq-dns-74dc88fc-sp4wt" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.527899 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kljb7\" (UniqueName: \"kubernetes.io/projected/43f1cabd-1a2d-4996-98e4-d90f58007fab-kube-api-access-kljb7\") pod \"dnsmasq-dns-74dc88fc-sp4wt\" (UID: \"43f1cabd-1a2d-4996-98e4-d90f58007fab\") " pod="openstack/dnsmasq-dns-74dc88fc-sp4wt" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.564634 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74dc88fc-sp4wt" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.942756 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-db-sync-ks59l" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.943365 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-db-sync-ks59l" event={"ID":"0ccef7a0-eebe-4794-b38d-690a98ee2055","Type":"ContainerDied","Data":"135419cd51312a93f574965953429b3364518e34b03dabcb160e9267c8a3877b"} Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.944436 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="135419cd51312a93f574965953429b3364518e34b03dabcb160e9267c8a3877b" Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.960553 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c09e3333-ad4d-495f-8973-7e2d2f0931d4","Type":"ContainerStarted","Data":"025102b1662bebf9881d4118f3dbbad014b7f2d347af8c5f1822d5c1d69e155d"} Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.960633 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c09e3333-ad4d-495f-8973-7e2d2f0931d4","Type":"ContainerStarted","Data":"548ab1edc33582fd93ac1f69d6e0f89d0a701ec7d069e9ef02800e0f08853d4b"} Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.960653 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c09e3333-ad4d-495f-8973-7e2d2f0931d4","Type":"ContainerStarted","Data":"6fdaf24b7a840946aee963464e5fb344d2a637db6e01251bf3d48bedee252cb2"} Oct 11 04:07:27 crc kubenswrapper[4967]: I1011 04:07:27.960672 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-storage-0" event={"ID":"c09e3333-ad4d-495f-8973-7e2d2f0931d4","Type":"ContainerStarted","Data":"f0cf0e9739ad65caa1198b66d7843e1888c89954d50472b971de0abf0eabf33f"} Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.016885 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-sp4wt"] Oct 11 04:07:28 crc kubenswrapper[4967]: W1011 04:07:28.017119 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod43f1cabd_1a2d_4996_98e4_d90f58007fab.slice/crio-a098b4defe0dce2c8aa25772beb0ba9723c3ad38587bbe2760ee25aa7f90c839 WatchSource:0}: Error finding container a098b4defe0dce2c8aa25772beb0ba9723c3ad38587bbe2760ee25aa7f90c839: Status 404 returned error can't find the container with id a098b4defe0dce2c8aa25772beb0ba9723c3ad38587bbe2760ee25aa7f90c839 Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.148264 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-storage-0" podStartSLOduration=43.391367761 podStartE2EDuration="52.148246001s" podCreationTimestamp="2025-10-11 04:06:36 +0000 UTC" firstStartedPulling="2025-10-11 04:07:17.606581294 +0000 UTC m=+965.569790227" lastFinishedPulling="2025-10-11 04:07:26.363459534 +0000 UTC m=+974.326668467" observedRunningTime="2025-10-11 04:07:28.02327349 +0000 UTC m=+975.986482453" watchObservedRunningTime="2025-10-11 04:07:28.148246001 +0000 UTC m=+976.111454934" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.157597 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-sp4wt"] Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.202086 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7d5679f497-nww24"] Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.203503 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d5679f497-nww24" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.235736 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-jmt45"] Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.236867 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jmt45" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.239012 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.239295 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-k7kxn" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.239513 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.239715 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.256755 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7d5679f497-nww24"] Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.262146 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-jmt45"] Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.333222 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31640236-7dcb-4bd8-851b-0b5870950b02-scripts\") pod \"keystone-bootstrap-jmt45\" (UID: \"31640236-7dcb-4bd8-851b-0b5870950b02\") " pod="openstack/keystone-bootstrap-jmt45" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.333273 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2nzdj\" (UniqueName: \"kubernetes.io/projected/31640236-7dcb-4bd8-851b-0b5870950b02-kube-api-access-2nzdj\") pod \"keystone-bootstrap-jmt45\" (UID: \"31640236-7dcb-4bd8-851b-0b5870950b02\") " pod="openstack/keystone-bootstrap-jmt45" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.333297 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31640236-7dcb-4bd8-851b-0b5870950b02-config-data\") pod \"keystone-bootstrap-jmt45\" (UID: \"31640236-7dcb-4bd8-851b-0b5870950b02\") " pod="openstack/keystone-bootstrap-jmt45" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.333318 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncwzv\" (UniqueName: \"kubernetes.io/projected/05507a50-8a11-44d2-94ea-5f72fe12a552-kube-api-access-ncwzv\") pod \"dnsmasq-dns-7d5679f497-nww24\" (UID: \"05507a50-8a11-44d2-94ea-5f72fe12a552\") " pod="openstack/dnsmasq-dns-7d5679f497-nww24" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.333340 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05507a50-8a11-44d2-94ea-5f72fe12a552-config\") pod \"dnsmasq-dns-7d5679f497-nww24\" (UID: \"05507a50-8a11-44d2-94ea-5f72fe12a552\") " pod="openstack/dnsmasq-dns-7d5679f497-nww24" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.333359 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/05507a50-8a11-44d2-94ea-5f72fe12a552-ovsdbserver-nb\") pod \"dnsmasq-dns-7d5679f497-nww24\" (UID: \"05507a50-8a11-44d2-94ea-5f72fe12a552\") " pod="openstack/dnsmasq-dns-7d5679f497-nww24" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.333378 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/31640236-7dcb-4bd8-851b-0b5870950b02-credential-keys\") pod \"keystone-bootstrap-jmt45\" (UID: \"31640236-7dcb-4bd8-851b-0b5870950b02\") " pod="openstack/keystone-bootstrap-jmt45" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.333402 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05507a50-8a11-44d2-94ea-5f72fe12a552-dns-svc\") pod \"dnsmasq-dns-7d5679f497-nww24\" (UID: \"05507a50-8a11-44d2-94ea-5f72fe12a552\") " pod="openstack/dnsmasq-dns-7d5679f497-nww24" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.333419 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31640236-7dcb-4bd8-851b-0b5870950b02-combined-ca-bundle\") pod \"keystone-bootstrap-jmt45\" (UID: \"31640236-7dcb-4bd8-851b-0b5870950b02\") " pod="openstack/keystone-bootstrap-jmt45" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.333484 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/05507a50-8a11-44d2-94ea-5f72fe12a552-ovsdbserver-sb\") pod \"dnsmasq-dns-7d5679f497-nww24\" (UID: \"05507a50-8a11-44d2-94ea-5f72fe12a552\") " pod="openstack/dnsmasq-dns-7d5679f497-nww24" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.333517 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/31640236-7dcb-4bd8-851b-0b5870950b02-fernet-keys\") pod \"keystone-bootstrap-jmt45\" (UID: \"31640236-7dcb-4bd8-851b-0b5870950b02\") " pod="openstack/keystone-bootstrap-jmt45" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.354705 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-db-create-4m9mw"] Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.357221 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-create-4m9mw" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.375976 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-db-create-4m9mw"] Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.429294 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d5679f497-nww24"] Oct 11 04:07:28 crc kubenswrapper[4967]: E1011 04:07:28.430170 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc kube-api-access-ncwzv ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-7d5679f497-nww24" podUID="05507a50-8a11-44d2-94ea-5f72fe12a552" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.439049 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/31640236-7dcb-4bd8-851b-0b5870950b02-fernet-keys\") pod \"keystone-bootstrap-jmt45\" (UID: \"31640236-7dcb-4bd8-851b-0b5870950b02\") " pod="openstack/keystone-bootstrap-jmt45" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.439124 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31640236-7dcb-4bd8-851b-0b5870950b02-scripts\") pod \"keystone-bootstrap-jmt45\" (UID: \"31640236-7dcb-4bd8-851b-0b5870950b02\") " pod="openstack/keystone-bootstrap-jmt45" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.439150 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2nzdj\" (UniqueName: \"kubernetes.io/projected/31640236-7dcb-4bd8-851b-0b5870950b02-kube-api-access-2nzdj\") pod \"keystone-bootstrap-jmt45\" (UID: \"31640236-7dcb-4bd8-851b-0b5870950b02\") " pod="openstack/keystone-bootstrap-jmt45" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.439192 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31640236-7dcb-4bd8-851b-0b5870950b02-config-data\") pod \"keystone-bootstrap-jmt45\" (UID: \"31640236-7dcb-4bd8-851b-0b5870950b02\") " pod="openstack/keystone-bootstrap-jmt45" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.439215 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncwzv\" (UniqueName: \"kubernetes.io/projected/05507a50-8a11-44d2-94ea-5f72fe12a552-kube-api-access-ncwzv\") pod \"dnsmasq-dns-7d5679f497-nww24\" (UID: \"05507a50-8a11-44d2-94ea-5f72fe12a552\") " pod="openstack/dnsmasq-dns-7d5679f497-nww24" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.439251 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05507a50-8a11-44d2-94ea-5f72fe12a552-config\") pod \"dnsmasq-dns-7d5679f497-nww24\" (UID: \"05507a50-8a11-44d2-94ea-5f72fe12a552\") " pod="openstack/dnsmasq-dns-7d5679f497-nww24" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.439272 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/05507a50-8a11-44d2-94ea-5f72fe12a552-ovsdbserver-nb\") pod \"dnsmasq-dns-7d5679f497-nww24\" (UID: \"05507a50-8a11-44d2-94ea-5f72fe12a552\") " pod="openstack/dnsmasq-dns-7d5679f497-nww24" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.439293 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/31640236-7dcb-4bd8-851b-0b5870950b02-credential-keys\") pod \"keystone-bootstrap-jmt45\" (UID: \"31640236-7dcb-4bd8-851b-0b5870950b02\") " pod="openstack/keystone-bootstrap-jmt45" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.439310 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05507a50-8a11-44d2-94ea-5f72fe12a552-dns-svc\") pod \"dnsmasq-dns-7d5679f497-nww24\" (UID: \"05507a50-8a11-44d2-94ea-5f72fe12a552\") " pod="openstack/dnsmasq-dns-7d5679f497-nww24" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.439349 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31640236-7dcb-4bd8-851b-0b5870950b02-combined-ca-bundle\") pod \"keystone-bootstrap-jmt45\" (UID: \"31640236-7dcb-4bd8-851b-0b5870950b02\") " pod="openstack/keystone-bootstrap-jmt45" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.439431 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/05507a50-8a11-44d2-94ea-5f72fe12a552-ovsdbserver-sb\") pod \"dnsmasq-dns-7d5679f497-nww24\" (UID: \"05507a50-8a11-44d2-94ea-5f72fe12a552\") " pod="openstack/dnsmasq-dns-7d5679f497-nww24" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.445874 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/05507a50-8a11-44d2-94ea-5f72fe12a552-ovsdbserver-sb\") pod \"dnsmasq-dns-7d5679f497-nww24\" (UID: \"05507a50-8a11-44d2-94ea-5f72fe12a552\") " pod="openstack/dnsmasq-dns-7d5679f497-nww24" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.446557 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05507a50-8a11-44d2-94ea-5f72fe12a552-dns-svc\") pod \"dnsmasq-dns-7d5679f497-nww24\" (UID: \"05507a50-8a11-44d2-94ea-5f72fe12a552\") " pod="openstack/dnsmasq-dns-7d5679f497-nww24" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.450235 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31640236-7dcb-4bd8-851b-0b5870950b02-combined-ca-bundle\") pod \"keystone-bootstrap-jmt45\" (UID: \"31640236-7dcb-4bd8-851b-0b5870950b02\") " pod="openstack/keystone-bootstrap-jmt45" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.450807 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05507a50-8a11-44d2-94ea-5f72fe12a552-config\") pod \"dnsmasq-dns-7d5679f497-nww24\" (UID: \"05507a50-8a11-44d2-94ea-5f72fe12a552\") " pod="openstack/dnsmasq-dns-7d5679f497-nww24" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.451704 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/05507a50-8a11-44d2-94ea-5f72fe12a552-ovsdbserver-nb\") pod \"dnsmasq-dns-7d5679f497-nww24\" (UID: \"05507a50-8a11-44d2-94ea-5f72fe12a552\") " pod="openstack/dnsmasq-dns-7d5679f497-nww24" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.451703 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/31640236-7dcb-4bd8-851b-0b5870950b02-fernet-keys\") pod \"keystone-bootstrap-jmt45\" (UID: \"31640236-7dcb-4bd8-851b-0b5870950b02\") " pod="openstack/keystone-bootstrap-jmt45" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.459062 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31640236-7dcb-4bd8-851b-0b5870950b02-scripts\") pod \"keystone-bootstrap-jmt45\" (UID: \"31640236-7dcb-4bd8-851b-0b5870950b02\") " pod="openstack/keystone-bootstrap-jmt45" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.460619 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31640236-7dcb-4bd8-851b-0b5870950b02-config-data\") pod \"keystone-bootstrap-jmt45\" (UID: \"31640236-7dcb-4bd8-851b-0b5870950b02\") " pod="openstack/keystone-bootstrap-jmt45" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.463778 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/31640236-7dcb-4bd8-851b-0b5870950b02-credential-keys\") pod \"keystone-bootstrap-jmt45\" (UID: \"31640236-7dcb-4bd8-851b-0b5870950b02\") " pod="openstack/keystone-bootstrap-jmt45" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.469298 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncwzv\" (UniqueName: \"kubernetes.io/projected/05507a50-8a11-44d2-94ea-5f72fe12a552-kube-api-access-ncwzv\") pod \"dnsmasq-dns-7d5679f497-nww24\" (UID: \"05507a50-8a11-44d2-94ea-5f72fe12a552\") " pod="openstack/dnsmasq-dns-7d5679f497-nww24" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.475229 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.477632 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2nzdj\" (UniqueName: \"kubernetes.io/projected/31640236-7dcb-4bd8-851b-0b5870950b02-kube-api-access-2nzdj\") pod \"keystone-bootstrap-jmt45\" (UID: \"31640236-7dcb-4bd8-851b-0b5870950b02\") " pod="openstack/keystone-bootstrap-jmt45" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.477843 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.488795 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.488986 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.497530 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-qbtj7"] Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.499148 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-qbtj7" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.501958 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"dns-swift-storage-0" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.506310 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.527628 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-qbtj7"] Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.548430 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c7sfz\" (UniqueName: \"kubernetes.io/projected/c70d7931-0ef5-44dc-88dd-b0f7b5ef3afd-kube-api-access-c7sfz\") pod \"ironic-db-create-4m9mw\" (UID: \"c70d7931-0ef5-44dc-88dd-b0f7b5ef3afd\") " pod="openstack/ironic-db-create-4m9mw" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.586960 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-qbtj7"] Oct 11 04:07:28 crc kubenswrapper[4967]: E1011 04:07:28.587676 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="unmounted volumes=[config dns-svc dns-swift-storage-0 kube-api-access-dcnsc ovsdbserver-nb ovsdbserver-sb], unattached volumes=[], failed to process volumes=[]: context canceled" pod="openstack/dnsmasq-dns-bbf5cc879-qbtj7" podUID="cdb41416-1288-433b-86d1-bfa03ac342c1" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.589322 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jmt45" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.609029 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-k8f6p"] Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.611641 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-k8f6p" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.620174 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-db-sync-zf6xs"] Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.621269 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zf6xs" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.623828 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-scripts" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.623993 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-config-data" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.624134 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"placement-placement-dockercfg-dcj9m" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.644242 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-k8f6p"] Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.644808 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-zf6xs"] Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.651579 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05059c3c-e29a-4bfa-a0d4-0d56b3d85af9-logs\") pod \"placement-db-sync-zf6xs\" (UID: \"05059c3c-e29a-4bfa-a0d4-0d56b3d85af9\") " pod="openstack/placement-db-sync-zf6xs" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.660194 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cdb41416-1288-433b-86d1-bfa03ac342c1-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-qbtj7\" (UID: \"cdb41416-1288-433b-86d1-bfa03ac342c1\") " pod="openstack/dnsmasq-dns-bbf5cc879-qbtj7" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.660497 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cdb41416-1288-433b-86d1-bfa03ac342c1-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-qbtj7\" (UID: \"cdb41416-1288-433b-86d1-bfa03ac342c1\") " pod="openstack/dnsmasq-dns-bbf5cc879-qbtj7" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.660593 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dfc995c0-dbf3-45a0-b16e-3d86acabe454-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-k8f6p\" (UID: \"dfc995c0-dbf3-45a0-b16e-3d86acabe454\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k8f6p" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.660686 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdb41416-1288-433b-86d1-bfa03ac342c1-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-qbtj7\" (UID: \"cdb41416-1288-433b-86d1-bfa03ac342c1\") " pod="openstack/dnsmasq-dns-bbf5cc879-qbtj7" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.660771 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05059c3c-e29a-4bfa-a0d4-0d56b3d85af9-config-data\") pod \"placement-db-sync-zf6xs\" (UID: \"05059c3c-e29a-4bfa-a0d4-0d56b3d85af9\") " pod="openstack/placement-db-sync-zf6xs" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.660888 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cdb41416-1288-433b-86d1-bfa03ac342c1-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-qbtj7\" (UID: \"cdb41416-1288-433b-86d1-bfa03ac342c1\") " pod="openstack/dnsmasq-dns-bbf5cc879-qbtj7" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.660994 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdb41416-1288-433b-86d1-bfa03ac342c1-config\") pod \"dnsmasq-dns-bbf5cc879-qbtj7\" (UID: \"cdb41416-1288-433b-86d1-bfa03ac342c1\") " pod="openstack/dnsmasq-dns-bbf5cc879-qbtj7" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.661095 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05059c3c-e29a-4bfa-a0d4-0d56b3d85af9-scripts\") pod \"placement-db-sync-zf6xs\" (UID: \"05059c3c-e29a-4bfa-a0d4-0d56b3d85af9\") " pod="openstack/placement-db-sync-zf6xs" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.661166 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kkrxv\" (UniqueName: \"kubernetes.io/projected/dfc995c0-dbf3-45a0-b16e-3d86acabe454-kube-api-access-kkrxv\") pod \"dnsmasq-dns-56df8fb6b7-k8f6p\" (UID: \"dfc995c0-dbf3-45a0-b16e-3d86acabe454\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k8f6p" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.661236 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dfc995c0-dbf3-45a0-b16e-3d86acabe454-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-k8f6p\" (UID: \"dfc995c0-dbf3-45a0-b16e-3d86acabe454\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k8f6p" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.661363 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcnsc\" (UniqueName: \"kubernetes.io/projected/cdb41416-1288-433b-86d1-bfa03ac342c1-kube-api-access-dcnsc\") pod \"dnsmasq-dns-bbf5cc879-qbtj7\" (UID: \"cdb41416-1288-433b-86d1-bfa03ac342c1\") " pod="openstack/dnsmasq-dns-bbf5cc879-qbtj7" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.661449 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05059c3c-e29a-4bfa-a0d4-0d56b3d85af9-combined-ca-bundle\") pod \"placement-db-sync-zf6xs\" (UID: \"05059c3c-e29a-4bfa-a0d4-0d56b3d85af9\") " pod="openstack/placement-db-sync-zf6xs" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.661546 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c7sfz\" (UniqueName: \"kubernetes.io/projected/c70d7931-0ef5-44dc-88dd-b0f7b5ef3afd-kube-api-access-c7sfz\") pod \"ironic-db-create-4m9mw\" (UID: \"c70d7931-0ef5-44dc-88dd-b0f7b5ef3afd\") " pod="openstack/ironic-db-create-4m9mw" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.661620 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w5r2v\" (UniqueName: \"kubernetes.io/projected/05059c3c-e29a-4bfa-a0d4-0d56b3d85af9-kube-api-access-w5r2v\") pod \"placement-db-sync-zf6xs\" (UID: \"05059c3c-e29a-4bfa-a0d4-0d56b3d85af9\") " pod="openstack/placement-db-sync-zf6xs" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.661701 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-scripts\") pod \"ceilometer-0\" (UID: \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\") " pod="openstack/ceilometer-0" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.661779 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-run-httpd\") pod \"ceilometer-0\" (UID: \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\") " pod="openstack/ceilometer-0" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.661871 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\") " pod="openstack/ceilometer-0" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.661943 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lfrp7\" (UniqueName: \"kubernetes.io/projected/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-kube-api-access-lfrp7\") pod \"ceilometer-0\" (UID: \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\") " pod="openstack/ceilometer-0" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.662021 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfc995c0-dbf3-45a0-b16e-3d86acabe454-config\") pod \"dnsmasq-dns-56df8fb6b7-k8f6p\" (UID: \"dfc995c0-dbf3-45a0-b16e-3d86acabe454\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k8f6p" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.662131 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dfc995c0-dbf3-45a0-b16e-3d86acabe454-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-k8f6p\" (UID: \"dfc995c0-dbf3-45a0-b16e-3d86acabe454\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k8f6p" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.662278 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-config-data\") pod \"ceilometer-0\" (UID: \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\") " pod="openstack/ceilometer-0" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.662416 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-log-httpd\") pod \"ceilometer-0\" (UID: \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\") " pod="openstack/ceilometer-0" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.662508 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dfc995c0-dbf3-45a0-b16e-3d86acabe454-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-k8f6p\" (UID: \"dfc995c0-dbf3-45a0-b16e-3d86acabe454\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k8f6p" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.662579 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\") " pod="openstack/ceilometer-0" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.695519 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c7sfz\" (UniqueName: \"kubernetes.io/projected/c70d7931-0ef5-44dc-88dd-b0f7b5ef3afd-kube-api-access-c7sfz\") pod \"ironic-db-create-4m9mw\" (UID: \"c70d7931-0ef5-44dc-88dd-b0f7b5ef3afd\") " pod="openstack/ironic-db-create-4m9mw" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.764542 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-log-httpd\") pod \"ceilometer-0\" (UID: \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\") " pod="openstack/ceilometer-0" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.764946 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dfc995c0-dbf3-45a0-b16e-3d86acabe454-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-k8f6p\" (UID: \"dfc995c0-dbf3-45a0-b16e-3d86acabe454\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k8f6p" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.764957 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-log-httpd\") pod \"ceilometer-0\" (UID: \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\") " pod="openstack/ceilometer-0" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.764975 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\") " pod="openstack/ceilometer-0" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.765127 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05059c3c-e29a-4bfa-a0d4-0d56b3d85af9-logs\") pod \"placement-db-sync-zf6xs\" (UID: \"05059c3c-e29a-4bfa-a0d4-0d56b3d85af9\") " pod="openstack/placement-db-sync-zf6xs" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.765158 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cdb41416-1288-433b-86d1-bfa03ac342c1-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-qbtj7\" (UID: \"cdb41416-1288-433b-86d1-bfa03ac342c1\") " pod="openstack/dnsmasq-dns-bbf5cc879-qbtj7" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.765197 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cdb41416-1288-433b-86d1-bfa03ac342c1-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-qbtj7\" (UID: \"cdb41416-1288-433b-86d1-bfa03ac342c1\") " pod="openstack/dnsmasq-dns-bbf5cc879-qbtj7" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.765222 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dfc995c0-dbf3-45a0-b16e-3d86acabe454-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-k8f6p\" (UID: \"dfc995c0-dbf3-45a0-b16e-3d86acabe454\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k8f6p" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.765253 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdb41416-1288-433b-86d1-bfa03ac342c1-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-qbtj7\" (UID: \"cdb41416-1288-433b-86d1-bfa03ac342c1\") " pod="openstack/dnsmasq-dns-bbf5cc879-qbtj7" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.765280 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cdb41416-1288-433b-86d1-bfa03ac342c1-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-qbtj7\" (UID: \"cdb41416-1288-433b-86d1-bfa03ac342c1\") " pod="openstack/dnsmasq-dns-bbf5cc879-qbtj7" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.765307 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05059c3c-e29a-4bfa-a0d4-0d56b3d85af9-config-data\") pod \"placement-db-sync-zf6xs\" (UID: \"05059c3c-e29a-4bfa-a0d4-0d56b3d85af9\") " pod="openstack/placement-db-sync-zf6xs" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.765340 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdb41416-1288-433b-86d1-bfa03ac342c1-config\") pod \"dnsmasq-dns-bbf5cc879-qbtj7\" (UID: \"cdb41416-1288-433b-86d1-bfa03ac342c1\") " pod="openstack/dnsmasq-dns-bbf5cc879-qbtj7" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.765364 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05059c3c-e29a-4bfa-a0d4-0d56b3d85af9-scripts\") pod \"placement-db-sync-zf6xs\" (UID: \"05059c3c-e29a-4bfa-a0d4-0d56b3d85af9\") " pod="openstack/placement-db-sync-zf6xs" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.765403 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kkrxv\" (UniqueName: \"kubernetes.io/projected/dfc995c0-dbf3-45a0-b16e-3d86acabe454-kube-api-access-kkrxv\") pod \"dnsmasq-dns-56df8fb6b7-k8f6p\" (UID: \"dfc995c0-dbf3-45a0-b16e-3d86acabe454\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k8f6p" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.765431 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dfc995c0-dbf3-45a0-b16e-3d86acabe454-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-k8f6p\" (UID: \"dfc995c0-dbf3-45a0-b16e-3d86acabe454\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k8f6p" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.765478 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcnsc\" (UniqueName: \"kubernetes.io/projected/cdb41416-1288-433b-86d1-bfa03ac342c1-kube-api-access-dcnsc\") pod \"dnsmasq-dns-bbf5cc879-qbtj7\" (UID: \"cdb41416-1288-433b-86d1-bfa03ac342c1\") " pod="openstack/dnsmasq-dns-bbf5cc879-qbtj7" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.765504 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05059c3c-e29a-4bfa-a0d4-0d56b3d85af9-combined-ca-bundle\") pod \"placement-db-sync-zf6xs\" (UID: \"05059c3c-e29a-4bfa-a0d4-0d56b3d85af9\") " pod="openstack/placement-db-sync-zf6xs" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.765540 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w5r2v\" (UniqueName: \"kubernetes.io/projected/05059c3c-e29a-4bfa-a0d4-0d56b3d85af9-kube-api-access-w5r2v\") pod \"placement-db-sync-zf6xs\" (UID: \"05059c3c-e29a-4bfa-a0d4-0d56b3d85af9\") " pod="openstack/placement-db-sync-zf6xs" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.765571 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-scripts\") pod \"ceilometer-0\" (UID: \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\") " pod="openstack/ceilometer-0" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.765587 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dfc995c0-dbf3-45a0-b16e-3d86acabe454-ovsdbserver-sb\") pod \"dnsmasq-dns-56df8fb6b7-k8f6p\" (UID: \"dfc995c0-dbf3-45a0-b16e-3d86acabe454\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k8f6p" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.765596 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-run-httpd\") pod \"ceilometer-0\" (UID: \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\") " pod="openstack/ceilometer-0" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.765632 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\") " pod="openstack/ceilometer-0" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.765660 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lfrp7\" (UniqueName: \"kubernetes.io/projected/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-kube-api-access-lfrp7\") pod \"ceilometer-0\" (UID: \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\") " pod="openstack/ceilometer-0" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.765690 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfc995c0-dbf3-45a0-b16e-3d86acabe454-config\") pod \"dnsmasq-dns-56df8fb6b7-k8f6p\" (UID: \"dfc995c0-dbf3-45a0-b16e-3d86acabe454\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k8f6p" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.765726 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dfc995c0-dbf3-45a0-b16e-3d86acabe454-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-k8f6p\" (UID: \"dfc995c0-dbf3-45a0-b16e-3d86acabe454\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k8f6p" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.765750 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-config-data\") pod \"ceilometer-0\" (UID: \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\") " pod="openstack/ceilometer-0" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.766482 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfc995c0-dbf3-45a0-b16e-3d86acabe454-config\") pod \"dnsmasq-dns-56df8fb6b7-k8f6p\" (UID: \"dfc995c0-dbf3-45a0-b16e-3d86acabe454\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k8f6p" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.766603 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cdb41416-1288-433b-86d1-bfa03ac342c1-dns-swift-storage-0\") pod \"dnsmasq-dns-bbf5cc879-qbtj7\" (UID: \"cdb41416-1288-433b-86d1-bfa03ac342c1\") " pod="openstack/dnsmasq-dns-bbf5cc879-qbtj7" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.766914 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdb41416-1288-433b-86d1-bfa03ac342c1-config\") pod \"dnsmasq-dns-bbf5cc879-qbtj7\" (UID: \"cdb41416-1288-433b-86d1-bfa03ac342c1\") " pod="openstack/dnsmasq-dns-bbf5cc879-qbtj7" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.767266 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dfc995c0-dbf3-45a0-b16e-3d86acabe454-ovsdbserver-nb\") pod \"dnsmasq-dns-56df8fb6b7-k8f6p\" (UID: \"dfc995c0-dbf3-45a0-b16e-3d86acabe454\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k8f6p" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.768596 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05059c3c-e29a-4bfa-a0d4-0d56b3d85af9-logs\") pod \"placement-db-sync-zf6xs\" (UID: \"05059c3c-e29a-4bfa-a0d4-0d56b3d85af9\") " pod="openstack/placement-db-sync-zf6xs" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.768709 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-run-httpd\") pod \"ceilometer-0\" (UID: \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\") " pod="openstack/ceilometer-0" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.769384 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdb41416-1288-433b-86d1-bfa03ac342c1-dns-svc\") pod \"dnsmasq-dns-bbf5cc879-qbtj7\" (UID: \"cdb41416-1288-433b-86d1-bfa03ac342c1\") " pod="openstack/dnsmasq-dns-bbf5cc879-qbtj7" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.769616 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cdb41416-1288-433b-86d1-bfa03ac342c1-ovsdbserver-nb\") pod \"dnsmasq-dns-bbf5cc879-qbtj7\" (UID: \"cdb41416-1288-433b-86d1-bfa03ac342c1\") " pod="openstack/dnsmasq-dns-bbf5cc879-qbtj7" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.769941 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cdb41416-1288-433b-86d1-bfa03ac342c1-ovsdbserver-sb\") pod \"dnsmasq-dns-bbf5cc879-qbtj7\" (UID: \"cdb41416-1288-433b-86d1-bfa03ac342c1\") " pod="openstack/dnsmasq-dns-bbf5cc879-qbtj7" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.771426 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dfc995c0-dbf3-45a0-b16e-3d86acabe454-dns-svc\") pod \"dnsmasq-dns-56df8fb6b7-k8f6p\" (UID: \"dfc995c0-dbf3-45a0-b16e-3d86acabe454\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k8f6p" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.773500 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dfc995c0-dbf3-45a0-b16e-3d86acabe454-dns-swift-storage-0\") pod \"dnsmasq-dns-56df8fb6b7-k8f6p\" (UID: \"dfc995c0-dbf3-45a0-b16e-3d86acabe454\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k8f6p" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.774510 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05059c3c-e29a-4bfa-a0d4-0d56b3d85af9-scripts\") pod \"placement-db-sync-zf6xs\" (UID: \"05059c3c-e29a-4bfa-a0d4-0d56b3d85af9\") " pod="openstack/placement-db-sync-zf6xs" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.774550 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\") " pod="openstack/ceilometer-0" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.780895 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-config-data\") pod \"ceilometer-0\" (UID: \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\") " pod="openstack/ceilometer-0" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.782964 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lfrp7\" (UniqueName: \"kubernetes.io/projected/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-kube-api-access-lfrp7\") pod \"ceilometer-0\" (UID: \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\") " pod="openstack/ceilometer-0" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.786019 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05059c3c-e29a-4bfa-a0d4-0d56b3d85af9-combined-ca-bundle\") pod \"placement-db-sync-zf6xs\" (UID: \"05059c3c-e29a-4bfa-a0d4-0d56b3d85af9\") " pod="openstack/placement-db-sync-zf6xs" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.790940 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-scripts\") pod \"ceilometer-0\" (UID: \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\") " pod="openstack/ceilometer-0" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.791470 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05059c3c-e29a-4bfa-a0d4-0d56b3d85af9-config-data\") pod \"placement-db-sync-zf6xs\" (UID: \"05059c3c-e29a-4bfa-a0d4-0d56b3d85af9\") " pod="openstack/placement-db-sync-zf6xs" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.791487 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\") " pod="openstack/ceilometer-0" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.797687 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcnsc\" (UniqueName: \"kubernetes.io/projected/cdb41416-1288-433b-86d1-bfa03ac342c1-kube-api-access-dcnsc\") pod \"dnsmasq-dns-bbf5cc879-qbtj7\" (UID: \"cdb41416-1288-433b-86d1-bfa03ac342c1\") " pod="openstack/dnsmasq-dns-bbf5cc879-qbtj7" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.804560 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w5r2v\" (UniqueName: \"kubernetes.io/projected/05059c3c-e29a-4bfa-a0d4-0d56b3d85af9-kube-api-access-w5r2v\") pod \"placement-db-sync-zf6xs\" (UID: \"05059c3c-e29a-4bfa-a0d4-0d56b3d85af9\") " pod="openstack/placement-db-sync-zf6xs" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.809728 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kkrxv\" (UniqueName: \"kubernetes.io/projected/dfc995c0-dbf3-45a0-b16e-3d86acabe454-kube-api-access-kkrxv\") pod \"dnsmasq-dns-56df8fb6b7-k8f6p\" (UID: \"dfc995c0-dbf3-45a0-b16e-3d86acabe454\") " pod="openstack/dnsmasq-dns-56df8fb6b7-k8f6p" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.821516 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-create-4m9mw" Oct 11 04:07:28 crc kubenswrapper[4967]: I1011 04:07:28.828873 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.036661 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-k8f6p" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.037238 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zf6xs" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.060666 4967 generic.go:334] "Generic (PLEG): container finished" podID="43f1cabd-1a2d-4996-98e4-d90f58007fab" containerID="ad2a4f1759205d2403cdb9000bed956dd64d73454b20cccfb83ed9be9d71db67" exitCode=0 Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.060754 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-qbtj7" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.061538 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74dc88fc-sp4wt" event={"ID":"43f1cabd-1a2d-4996-98e4-d90f58007fab","Type":"ContainerDied","Data":"ad2a4f1759205d2403cdb9000bed956dd64d73454b20cccfb83ed9be9d71db67"} Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.061562 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74dc88fc-sp4wt" event={"ID":"43f1cabd-1a2d-4996-98e4-d90f58007fab","Type":"ContainerStarted","Data":"a098b4defe0dce2c8aa25772beb0ba9723c3ad38587bbe2760ee25aa7f90c839"} Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.062708 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d5679f497-nww24" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.077093 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d5679f497-nww24" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.077283 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-qbtj7" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.148634 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-jmt45"] Oct 11 04:07:29 crc kubenswrapper[4967]: W1011 04:07:29.167968 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod31640236_7dcb_4bd8_851b_0b5870950b02.slice/crio-29b3adc73ac1f1ac2e7537e5f0d514f3884577e5da05b844bde9d925e62fa6bc WatchSource:0}: Error finding container 29b3adc73ac1f1ac2e7537e5f0d514f3884577e5da05b844bde9d925e62fa6bc: Status 404 returned error can't find the container with id 29b3adc73ac1f1ac2e7537e5f0d514f3884577e5da05b844bde9d925e62fa6bc Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.281024 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/05507a50-8a11-44d2-94ea-5f72fe12a552-ovsdbserver-nb\") pod \"05507a50-8a11-44d2-94ea-5f72fe12a552\" (UID: \"05507a50-8a11-44d2-94ea-5f72fe12a552\") " Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.281085 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cdb41416-1288-433b-86d1-bfa03ac342c1-ovsdbserver-nb\") pod \"cdb41416-1288-433b-86d1-bfa03ac342c1\" (UID: \"cdb41416-1288-433b-86d1-bfa03ac342c1\") " Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.281126 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cdb41416-1288-433b-86d1-bfa03ac342c1-dns-swift-storage-0\") pod \"cdb41416-1288-433b-86d1-bfa03ac342c1\" (UID: \"cdb41416-1288-433b-86d1-bfa03ac342c1\") " Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.281150 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05507a50-8a11-44d2-94ea-5f72fe12a552-dns-svc\") pod \"05507a50-8a11-44d2-94ea-5f72fe12a552\" (UID: \"05507a50-8a11-44d2-94ea-5f72fe12a552\") " Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.281172 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cdb41416-1288-433b-86d1-bfa03ac342c1-ovsdbserver-sb\") pod \"cdb41416-1288-433b-86d1-bfa03ac342c1\" (UID: \"cdb41416-1288-433b-86d1-bfa03ac342c1\") " Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.281193 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdb41416-1288-433b-86d1-bfa03ac342c1-dns-svc\") pod \"cdb41416-1288-433b-86d1-bfa03ac342c1\" (UID: \"cdb41416-1288-433b-86d1-bfa03ac342c1\") " Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.281227 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdb41416-1288-433b-86d1-bfa03ac342c1-config\") pod \"cdb41416-1288-433b-86d1-bfa03ac342c1\" (UID: \"cdb41416-1288-433b-86d1-bfa03ac342c1\") " Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.281264 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ncwzv\" (UniqueName: \"kubernetes.io/projected/05507a50-8a11-44d2-94ea-5f72fe12a552-kube-api-access-ncwzv\") pod \"05507a50-8a11-44d2-94ea-5f72fe12a552\" (UID: \"05507a50-8a11-44d2-94ea-5f72fe12a552\") " Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.281299 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcnsc\" (UniqueName: \"kubernetes.io/projected/cdb41416-1288-433b-86d1-bfa03ac342c1-kube-api-access-dcnsc\") pod \"cdb41416-1288-433b-86d1-bfa03ac342c1\" (UID: \"cdb41416-1288-433b-86d1-bfa03ac342c1\") " Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.281326 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05507a50-8a11-44d2-94ea-5f72fe12a552-config\") pod \"05507a50-8a11-44d2-94ea-5f72fe12a552\" (UID: \"05507a50-8a11-44d2-94ea-5f72fe12a552\") " Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.281443 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/05507a50-8a11-44d2-94ea-5f72fe12a552-ovsdbserver-sb\") pod \"05507a50-8a11-44d2-94ea-5f72fe12a552\" (UID: \"05507a50-8a11-44d2-94ea-5f72fe12a552\") " Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.281628 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05507a50-8a11-44d2-94ea-5f72fe12a552-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "05507a50-8a11-44d2-94ea-5f72fe12a552" (UID: "05507a50-8a11-44d2-94ea-5f72fe12a552"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.282358 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdb41416-1288-433b-86d1-bfa03ac342c1-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "cdb41416-1288-433b-86d1-bfa03ac342c1" (UID: "cdb41416-1288-433b-86d1-bfa03ac342c1"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.282442 4967 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/05507a50-8a11-44d2-94ea-5f72fe12a552-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.282783 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdb41416-1288-433b-86d1-bfa03ac342c1-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "cdb41416-1288-433b-86d1-bfa03ac342c1" (UID: "cdb41416-1288-433b-86d1-bfa03ac342c1"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.283150 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05507a50-8a11-44d2-94ea-5f72fe12a552-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "05507a50-8a11-44d2-94ea-5f72fe12a552" (UID: "05507a50-8a11-44d2-94ea-5f72fe12a552"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.283341 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05507a50-8a11-44d2-94ea-5f72fe12a552-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "05507a50-8a11-44d2-94ea-5f72fe12a552" (UID: "05507a50-8a11-44d2-94ea-5f72fe12a552"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.283474 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdb41416-1288-433b-86d1-bfa03ac342c1-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "cdb41416-1288-433b-86d1-bfa03ac342c1" (UID: "cdb41416-1288-433b-86d1-bfa03ac342c1"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.283790 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/05507a50-8a11-44d2-94ea-5f72fe12a552-config" (OuterVolumeSpecName: "config") pod "05507a50-8a11-44d2-94ea-5f72fe12a552" (UID: "05507a50-8a11-44d2-94ea-5f72fe12a552"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.284144 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdb41416-1288-433b-86d1-bfa03ac342c1-config" (OuterVolumeSpecName: "config") pod "cdb41416-1288-433b-86d1-bfa03ac342c1" (UID: "cdb41416-1288-433b-86d1-bfa03ac342c1"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.284770 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cdb41416-1288-433b-86d1-bfa03ac342c1-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "cdb41416-1288-433b-86d1-bfa03ac342c1" (UID: "cdb41416-1288-433b-86d1-bfa03ac342c1"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.287652 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cdb41416-1288-433b-86d1-bfa03ac342c1-kube-api-access-dcnsc" (OuterVolumeSpecName: "kube-api-access-dcnsc") pod "cdb41416-1288-433b-86d1-bfa03ac342c1" (UID: "cdb41416-1288-433b-86d1-bfa03ac342c1"). InnerVolumeSpecName "kube-api-access-dcnsc". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.287880 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05507a50-8a11-44d2-94ea-5f72fe12a552-kube-api-access-ncwzv" (OuterVolumeSpecName: "kube-api-access-ncwzv") pod "05507a50-8a11-44d2-94ea-5f72fe12a552" (UID: "05507a50-8a11-44d2-94ea-5f72fe12a552"). InnerVolumeSpecName "kube-api-access-ncwzv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.346082 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.350240 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.352025 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-2d8h2" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.352276 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.352452 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.360953 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.384224 4967 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/05507a50-8a11-44d2-94ea-5f72fe12a552-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.384254 4967 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/cdb41416-1288-433b-86d1-bfa03ac342c1-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.384263 4967 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/cdb41416-1288-433b-86d1-bfa03ac342c1-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.384272 4967 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/05507a50-8a11-44d2-94ea-5f72fe12a552-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.384281 4967 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/cdb41416-1288-433b-86d1-bfa03ac342c1-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.384291 4967 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/cdb41416-1288-433b-86d1-bfa03ac342c1-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.384298 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cdb41416-1288-433b-86d1-bfa03ac342c1-config\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.384306 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ncwzv\" (UniqueName: \"kubernetes.io/projected/05507a50-8a11-44d2-94ea-5f72fe12a552-kube-api-access-ncwzv\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.384316 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcnsc\" (UniqueName: \"kubernetes.io/projected/cdb41416-1288-433b-86d1-bfa03ac342c1-kube-api-access-dcnsc\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.384326 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/05507a50-8a11-44d2-94ea-5f72fe12a552-config\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:29 crc kubenswrapper[4967]: W1011 04:07:29.476465 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc70d7931_0ef5_44dc_88dd_b0f7b5ef3afd.slice/crio-52fc00d0e5c3ce9a5e134afed92ae6c487e1216b32cc0103a2372e6567e4a69b WatchSource:0}: Error finding container 52fc00d0e5c3ce9a5e134afed92ae6c487e1216b32cc0103a2372e6567e4a69b: Status 404 returned error can't find the container with id 52fc00d0e5c3ce9a5e134afed92ae6c487e1216b32cc0103a2372e6567e4a69b Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.485122 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ddabca8-7c71-431e-b3f6-987d21900b0d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3ddabca8-7c71-431e-b3f6-987d21900b0d\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.485161 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3ddabca8-7c71-431e-b3f6-987d21900b0d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3ddabca8-7c71-431e-b3f6-987d21900b0d\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.485179 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zq7mz\" (UniqueName: \"kubernetes.io/projected/3ddabca8-7c71-431e-b3f6-987d21900b0d-kube-api-access-zq7mz\") pod \"glance-default-external-api-0\" (UID: \"3ddabca8-7c71-431e-b3f6-987d21900b0d\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.485215 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ddabca8-7c71-431e-b3f6-987d21900b0d-config-data\") pod \"glance-default-external-api-0\" (UID: \"3ddabca8-7c71-431e-b3f6-987d21900b0d\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.485388 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ddabca8-7c71-431e-b3f6-987d21900b0d-scripts\") pod \"glance-default-external-api-0\" (UID: \"3ddabca8-7c71-431e-b3f6-987d21900b0d\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.485427 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ddabca8-7c71-431e-b3f6-987d21900b0d-logs\") pod \"glance-default-external-api-0\" (UID: \"3ddabca8-7c71-431e-b3f6-987d21900b0d\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.485492 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"3ddabca8-7c71-431e-b3f6-987d21900b0d\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.489427 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-db-create-4m9mw"] Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.510957 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.512588 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.517617 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.526662 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74dc88fc-sp4wt" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.529536 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.587673 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ddabca8-7c71-431e-b3f6-987d21900b0d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3ddabca8-7c71-431e-b3f6-987d21900b0d\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.587727 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3ddabca8-7c71-431e-b3f6-987d21900b0d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3ddabca8-7c71-431e-b3f6-987d21900b0d\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.587759 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zq7mz\" (UniqueName: \"kubernetes.io/projected/3ddabca8-7c71-431e-b3f6-987d21900b0d-kube-api-access-zq7mz\") pod \"glance-default-external-api-0\" (UID: \"3ddabca8-7c71-431e-b3f6-987d21900b0d\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.587820 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ddabca8-7c71-431e-b3f6-987d21900b0d-config-data\") pod \"glance-default-external-api-0\" (UID: \"3ddabca8-7c71-431e-b3f6-987d21900b0d\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.587887 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ddabca8-7c71-431e-b3f6-987d21900b0d-scripts\") pod \"glance-default-external-api-0\" (UID: \"3ddabca8-7c71-431e-b3f6-987d21900b0d\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.587933 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ddabca8-7c71-431e-b3f6-987d21900b0d-logs\") pod \"glance-default-external-api-0\" (UID: \"3ddabca8-7c71-431e-b3f6-987d21900b0d\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.588000 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"3ddabca8-7c71-431e-b3f6-987d21900b0d\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.588435 4967 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"3ddabca8-7c71-431e-b3f6-987d21900b0d\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.588934 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3ddabca8-7c71-431e-b3f6-987d21900b0d-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"3ddabca8-7c71-431e-b3f6-987d21900b0d\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.589438 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ddabca8-7c71-431e-b3f6-987d21900b0d-logs\") pod \"glance-default-external-api-0\" (UID: \"3ddabca8-7c71-431e-b3f6-987d21900b0d\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.592944 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ddabca8-7c71-431e-b3f6-987d21900b0d-config-data\") pod \"glance-default-external-api-0\" (UID: \"3ddabca8-7c71-431e-b3f6-987d21900b0d\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.593695 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ddabca8-7c71-431e-b3f6-987d21900b0d-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"3ddabca8-7c71-431e-b3f6-987d21900b0d\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.613005 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ddabca8-7c71-431e-b3f6-987d21900b0d-scripts\") pod \"glance-default-external-api-0\" (UID: \"3ddabca8-7c71-431e-b3f6-987d21900b0d\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.619861 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zq7mz\" (UniqueName: \"kubernetes.io/projected/3ddabca8-7c71-431e-b3f6-987d21900b0d-kube-api-access-zq7mz\") pod \"glance-default-external-api-0\" (UID: \"3ddabca8-7c71-431e-b3f6-987d21900b0d\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.621982 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"3ddabca8-7c71-431e-b3f6-987d21900b0d\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.640880 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:07:29 crc kubenswrapper[4967]: W1011 04:07:29.644310 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podeb56ccd5_8cd3_4da0_b11f_8600af8abf86.slice/crio-a8f828a9cd7f97607aa052ce81010e0423cf521eec734e13ca4ea8d5fb4fa18d WatchSource:0}: Error finding container a8f828a9cd7f97607aa052ce81010e0423cf521eec734e13ca4ea8d5fb4fa18d: Status 404 returned error can't find the container with id a8f828a9cd7f97607aa052ce81010e0423cf521eec734e13ca4ea8d5fb4fa18d Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.689524 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/43f1cabd-1a2d-4996-98e4-d90f58007fab-ovsdbserver-nb\") pod \"43f1cabd-1a2d-4996-98e4-d90f58007fab\" (UID: \"43f1cabd-1a2d-4996-98e4-d90f58007fab\") " Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.689805 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43f1cabd-1a2d-4996-98e4-d90f58007fab-dns-svc\") pod \"43f1cabd-1a2d-4996-98e4-d90f58007fab\" (UID: \"43f1cabd-1a2d-4996-98e4-d90f58007fab\") " Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.689830 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kljb7\" (UniqueName: \"kubernetes.io/projected/43f1cabd-1a2d-4996-98e4-d90f58007fab-kube-api-access-kljb7\") pod \"43f1cabd-1a2d-4996-98e4-d90f58007fab\" (UID: \"43f1cabd-1a2d-4996-98e4-d90f58007fab\") " Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.689875 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43f1cabd-1a2d-4996-98e4-d90f58007fab-config\") pod \"43f1cabd-1a2d-4996-98e4-d90f58007fab\" (UID: \"43f1cabd-1a2d-4996-98e4-d90f58007fab\") " Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.689905 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/43f1cabd-1a2d-4996-98e4-d90f58007fab-ovsdbserver-sb\") pod \"43f1cabd-1a2d-4996-98e4-d90f58007fab\" (UID: \"43f1cabd-1a2d-4996-98e4-d90f58007fab\") " Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.690094 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.690123 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ss5pf\" (UniqueName: \"kubernetes.io/projected/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-kube-api-access-ss5pf\") pod \"glance-default-internal-api-0\" (UID: \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.690147 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.690186 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.690209 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.690259 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-logs\") pod \"glance-default-internal-api-0\" (UID: \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.690274 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.693857 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43f1cabd-1a2d-4996-98e4-d90f58007fab-kube-api-access-kljb7" (OuterVolumeSpecName: "kube-api-access-kljb7") pod "43f1cabd-1a2d-4996-98e4-d90f58007fab" (UID: "43f1cabd-1a2d-4996-98e4-d90f58007fab"). InnerVolumeSpecName "kube-api-access-kljb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.697206 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.716524 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43f1cabd-1a2d-4996-98e4-d90f58007fab-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "43f1cabd-1a2d-4996-98e4-d90f58007fab" (UID: "43f1cabd-1a2d-4996-98e4-d90f58007fab"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.718065 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43f1cabd-1a2d-4996-98e4-d90f58007fab-config" (OuterVolumeSpecName: "config") pod "43f1cabd-1a2d-4996-98e4-d90f58007fab" (UID: "43f1cabd-1a2d-4996-98e4-d90f58007fab"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.720103 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43f1cabd-1a2d-4996-98e4-d90f58007fab-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "43f1cabd-1a2d-4996-98e4-d90f58007fab" (UID: "43f1cabd-1a2d-4996-98e4-d90f58007fab"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.721140 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43f1cabd-1a2d-4996-98e4-d90f58007fab-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "43f1cabd-1a2d-4996-98e4-d90f58007fab" (UID: "43f1cabd-1a2d-4996-98e4-d90f58007fab"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.763321 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-db-sync-zf6xs"] Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.768834 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-k8f6p"] Oct 11 04:07:29 crc kubenswrapper[4967]: W1011 04:07:29.773395 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddfc995c0_dbf3_45a0_b16e_3d86acabe454.slice/crio-b7edf5c52368d51a6e2f60b140c3b33189073dc358e1d1a4bc9494d28329c453 WatchSource:0}: Error finding container b7edf5c52368d51a6e2f60b140c3b33189073dc358e1d1a4bc9494d28329c453: Status 404 returned error can't find the container with id b7edf5c52368d51a6e2f60b140c3b33189073dc358e1d1a4bc9494d28329c453 Oct 11 04:07:29 crc kubenswrapper[4967]: W1011 04:07:29.780344 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod05059c3c_e29a_4bfa_a0d4_0d56b3d85af9.slice/crio-2174c3ec42975523ddd78c08814d157c30d6a60c2699ff360bf600e540168733 WatchSource:0}: Error finding container 2174c3ec42975523ddd78c08814d157c30d6a60c2699ff360bf600e540168733: Status 404 returned error can't find the container with id 2174c3ec42975523ddd78c08814d157c30d6a60c2699ff360bf600e540168733 Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.791922 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.791962 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.792024 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-logs\") pod \"glance-default-internal-api-0\" (UID: \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.792040 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.792100 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.792126 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ss5pf\" (UniqueName: \"kubernetes.io/projected/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-kube-api-access-ss5pf\") pod \"glance-default-internal-api-0\" (UID: \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.792149 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.792190 4967 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/43f1cabd-1a2d-4996-98e4-d90f58007fab-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.792202 4967 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/43f1cabd-1a2d-4996-98e4-d90f58007fab-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.792211 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kljb7\" (UniqueName: \"kubernetes.io/projected/43f1cabd-1a2d-4996-98e4-d90f58007fab-kube-api-access-kljb7\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.792221 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/43f1cabd-1a2d-4996-98e4-d90f58007fab-config\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.792229 4967 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/43f1cabd-1a2d-4996-98e4-d90f58007fab-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.793351 4967 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.794411 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-logs\") pod \"glance-default-internal-api-0\" (UID: \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.794728 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.802437 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-scripts\") pod \"glance-default-internal-api-0\" (UID: \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.803379 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-config-data\") pod \"glance-default-internal-api-0\" (UID: \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.824158 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ss5pf\" (UniqueName: \"kubernetes.io/projected/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-kube-api-access-ss5pf\") pod \"glance-default-internal-api-0\" (UID: \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.826285 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.834243 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:29 crc kubenswrapper[4967]: I1011 04:07:29.933486 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 11 04:07:30 crc kubenswrapper[4967]: I1011 04:07:30.071030 4967 generic.go:334] "Generic (PLEG): container finished" podID="dfc995c0-dbf3-45a0-b16e-3d86acabe454" containerID="7fb30d62397b4f0465dd75c26bfec185771fabfeee3c3279a171d97a3ffa5e8c" exitCode=0 Oct 11 04:07:30 crc kubenswrapper[4967]: I1011 04:07:30.071105 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-k8f6p" event={"ID":"dfc995c0-dbf3-45a0-b16e-3d86acabe454","Type":"ContainerDied","Data":"7fb30d62397b4f0465dd75c26bfec185771fabfeee3c3279a171d97a3ffa5e8c"} Oct 11 04:07:30 crc kubenswrapper[4967]: I1011 04:07:30.071156 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-k8f6p" event={"ID":"dfc995c0-dbf3-45a0-b16e-3d86acabe454","Type":"ContainerStarted","Data":"b7edf5c52368d51a6e2f60b140c3b33189073dc358e1d1a4bc9494d28329c453"} Oct 11 04:07:30 crc kubenswrapper[4967]: I1011 04:07:30.074793 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zf6xs" event={"ID":"05059c3c-e29a-4bfa-a0d4-0d56b3d85af9","Type":"ContainerStarted","Data":"2174c3ec42975523ddd78c08814d157c30d6a60c2699ff360bf600e540168733"} Oct 11 04:07:30 crc kubenswrapper[4967]: I1011 04:07:30.077848 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jmt45" event={"ID":"31640236-7dcb-4bd8-851b-0b5870950b02","Type":"ContainerStarted","Data":"8d452e5bcdee06ff27ad299bd16449ebf91079000282280714301c18395e32a6"} Oct 11 04:07:30 crc kubenswrapper[4967]: I1011 04:07:30.077884 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jmt45" event={"ID":"31640236-7dcb-4bd8-851b-0b5870950b02","Type":"ContainerStarted","Data":"29b3adc73ac1f1ac2e7537e5f0d514f3884577e5da05b844bde9d925e62fa6bc"} Oct 11 04:07:30 crc kubenswrapper[4967]: I1011 04:07:30.080537 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-74dc88fc-sp4wt" event={"ID":"43f1cabd-1a2d-4996-98e4-d90f58007fab","Type":"ContainerDied","Data":"a098b4defe0dce2c8aa25772beb0ba9723c3ad38587bbe2760ee25aa7f90c839"} Oct 11 04:07:30 crc kubenswrapper[4967]: I1011 04:07:30.080575 4967 scope.go:117] "RemoveContainer" containerID="ad2a4f1759205d2403cdb9000bed956dd64d73454b20cccfb83ed9be9d71db67" Oct 11 04:07:30 crc kubenswrapper[4967]: I1011 04:07:30.080692 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-74dc88fc-sp4wt" Oct 11 04:07:30 crc kubenswrapper[4967]: I1011 04:07:30.090893 4967 generic.go:334] "Generic (PLEG): container finished" podID="c70d7931-0ef5-44dc-88dd-b0f7b5ef3afd" containerID="240ba3af22688386cb63250f6b108ce87c2b90ac426742afa8ae2ed6c30da0cd" exitCode=0 Oct 11 04:07:30 crc kubenswrapper[4967]: I1011 04:07:30.091040 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-create-4m9mw" event={"ID":"c70d7931-0ef5-44dc-88dd-b0f7b5ef3afd","Type":"ContainerDied","Data":"240ba3af22688386cb63250f6b108ce87c2b90ac426742afa8ae2ed6c30da0cd"} Oct 11 04:07:30 crc kubenswrapper[4967]: I1011 04:07:30.091118 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-create-4m9mw" event={"ID":"c70d7931-0ef5-44dc-88dd-b0f7b5ef3afd","Type":"ContainerStarted","Data":"52fc00d0e5c3ce9a5e134afed92ae6c487e1216b32cc0103a2372e6567e4a69b"} Oct 11 04:07:30 crc kubenswrapper[4967]: I1011 04:07:30.102508 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb56ccd5-8cd3-4da0-b11f-8600af8abf86","Type":"ContainerStarted","Data":"a8f828a9cd7f97607aa052ce81010e0423cf521eec734e13ca4ea8d5fb4fa18d"} Oct 11 04:07:30 crc kubenswrapper[4967]: I1011 04:07:30.102812 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bbf5cc879-qbtj7" Oct 11 04:07:30 crc kubenswrapper[4967]: I1011 04:07:30.110711 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-jmt45" podStartSLOduration=2.110693936 podStartE2EDuration="2.110693936s" podCreationTimestamp="2025-10-11 04:07:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:07:30.109269533 +0000 UTC m=+978.072478466" watchObservedRunningTime="2025-10-11 04:07:30.110693936 +0000 UTC m=+978.073902869" Oct 11 04:07:30 crc kubenswrapper[4967]: I1011 04:07:30.123773 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7d5679f497-nww24" Oct 11 04:07:30 crc kubenswrapper[4967]: I1011 04:07:30.184957 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-sp4wt"] Oct 11 04:07:30 crc kubenswrapper[4967]: I1011 04:07:30.231358 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-74dc88fc-sp4wt"] Oct 11 04:07:30 crc kubenswrapper[4967]: I1011 04:07:30.259862 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-qbtj7"] Oct 11 04:07:30 crc kubenswrapper[4967]: I1011 04:07:30.266034 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bbf5cc879-qbtj7"] Oct 11 04:07:30 crc kubenswrapper[4967]: I1011 04:07:30.281778 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7d5679f497-nww24"] Oct 11 04:07:30 crc kubenswrapper[4967]: I1011 04:07:30.287829 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7d5679f497-nww24"] Oct 11 04:07:30 crc kubenswrapper[4967]: I1011 04:07:30.292914 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 11 04:07:30 crc kubenswrapper[4967]: I1011 04:07:30.548522 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 11 04:07:30 crc kubenswrapper[4967]: I1011 04:07:30.865051 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05507a50-8a11-44d2-94ea-5f72fe12a552" path="/var/lib/kubelet/pods/05507a50-8a11-44d2-94ea-5f72fe12a552/volumes" Oct 11 04:07:30 crc kubenswrapper[4967]: I1011 04:07:30.865497 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43f1cabd-1a2d-4996-98e4-d90f58007fab" path="/var/lib/kubelet/pods/43f1cabd-1a2d-4996-98e4-d90f58007fab/volumes" Oct 11 04:07:30 crc kubenswrapper[4967]: I1011 04:07:30.866255 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cdb41416-1288-433b-86d1-bfa03ac342c1" path="/var/lib/kubelet/pods/cdb41416-1288-433b-86d1-bfa03ac342c1/volumes" Oct 11 04:07:30 crc kubenswrapper[4967]: I1011 04:07:30.866574 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 11 04:07:30 crc kubenswrapper[4967]: I1011 04:07:30.866601 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:07:30 crc kubenswrapper[4967]: I1011 04:07:30.896747 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 11 04:07:31 crc kubenswrapper[4967]: I1011 04:07:31.121222 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3ddabca8-7c71-431e-b3f6-987d21900b0d","Type":"ContainerStarted","Data":"7fc00872eb20d715869841e872d53e5203e2c809679a6cf3edc55987d07c4441"} Oct 11 04:07:31 crc kubenswrapper[4967]: I1011 04:07:31.121577 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3ddabca8-7c71-431e-b3f6-987d21900b0d","Type":"ContainerStarted","Data":"e6b952693f5a79981f71f9a4424a7e1b7b1443dff3d37e49cd3bee4a22ae1e68"} Oct 11 04:07:31 crc kubenswrapper[4967]: I1011 04:07:31.137455 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-k8f6p" event={"ID":"dfc995c0-dbf3-45a0-b16e-3d86acabe454","Type":"ContainerStarted","Data":"1853439d5d6a6d705b66fdd687606d349872991825f0476ad70d3f6f597aded2"} Oct 11 04:07:31 crc kubenswrapper[4967]: I1011 04:07:31.137654 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-56df8fb6b7-k8f6p" Oct 11 04:07:31 crc kubenswrapper[4967]: I1011 04:07:31.139939 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b92605b2-50a6-41bc-8a34-8783cc9fc7c9","Type":"ContainerStarted","Data":"a8c9b656af8d8321b27ecdb510d6f332ff15bd517bb4746fab1589a3736a3fdd"} Oct 11 04:07:31 crc kubenswrapper[4967]: I1011 04:07:31.156981 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-56df8fb6b7-k8f6p" podStartSLOduration=3.156965183 podStartE2EDuration="3.156965183s" podCreationTimestamp="2025-10-11 04:07:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:07:31.151839525 +0000 UTC m=+979.115048458" watchObservedRunningTime="2025-10-11 04:07:31.156965183 +0000 UTC m=+979.120174116" Oct 11 04:07:31 crc kubenswrapper[4967]: I1011 04:07:31.501753 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-create-4m9mw" Oct 11 04:07:31 crc kubenswrapper[4967]: I1011 04:07:31.639182 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c7sfz\" (UniqueName: \"kubernetes.io/projected/c70d7931-0ef5-44dc-88dd-b0f7b5ef3afd-kube-api-access-c7sfz\") pod \"c70d7931-0ef5-44dc-88dd-b0f7b5ef3afd\" (UID: \"c70d7931-0ef5-44dc-88dd-b0f7b5ef3afd\") " Oct 11 04:07:31 crc kubenswrapper[4967]: I1011 04:07:31.644557 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c70d7931-0ef5-44dc-88dd-b0f7b5ef3afd-kube-api-access-c7sfz" (OuterVolumeSpecName: "kube-api-access-c7sfz") pod "c70d7931-0ef5-44dc-88dd-b0f7b5ef3afd" (UID: "c70d7931-0ef5-44dc-88dd-b0f7b5ef3afd"). InnerVolumeSpecName "kube-api-access-c7sfz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:07:31 crc kubenswrapper[4967]: I1011 04:07:31.741264 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c7sfz\" (UniqueName: \"kubernetes.io/projected/c70d7931-0ef5-44dc-88dd-b0f7b5ef3afd-kube-api-access-c7sfz\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:32 crc kubenswrapper[4967]: I1011 04:07:32.155047 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b92605b2-50a6-41bc-8a34-8783cc9fc7c9","Type":"ContainerStarted","Data":"dea40148c9a1a67da3b4bd3d86f7758d94d9b0d3207421731cb4c1e6dbc02697"} Oct 11 04:07:32 crc kubenswrapper[4967]: I1011 04:07:32.155533 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b92605b2-50a6-41bc-8a34-8783cc9fc7c9","Type":"ContainerStarted","Data":"e06f2ac89aa25dc24dc2dab0cb337b7fe9eafd14405c4942ed856a86f494bb9d"} Oct 11 04:07:32 crc kubenswrapper[4967]: I1011 04:07:32.155646 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="b92605b2-50a6-41bc-8a34-8783cc9fc7c9" containerName="glance-log" containerID="cri-o://e06f2ac89aa25dc24dc2dab0cb337b7fe9eafd14405c4942ed856a86f494bb9d" gracePeriod=30 Oct 11 04:07:32 crc kubenswrapper[4967]: I1011 04:07:32.156017 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="b92605b2-50a6-41bc-8a34-8783cc9fc7c9" containerName="glance-httpd" containerID="cri-o://dea40148c9a1a67da3b4bd3d86f7758d94d9b0d3207421731cb4c1e6dbc02697" gracePeriod=30 Oct 11 04:07:32 crc kubenswrapper[4967]: I1011 04:07:32.159619 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-create-4m9mw" event={"ID":"c70d7931-0ef5-44dc-88dd-b0f7b5ef3afd","Type":"ContainerDied","Data":"52fc00d0e5c3ce9a5e134afed92ae6c487e1216b32cc0103a2372e6567e4a69b"} Oct 11 04:07:32 crc kubenswrapper[4967]: I1011 04:07:32.159654 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="52fc00d0e5c3ce9a5e134afed92ae6c487e1216b32cc0103a2372e6567e4a69b" Oct 11 04:07:32 crc kubenswrapper[4967]: I1011 04:07:32.159708 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-create-4m9mw" Oct 11 04:07:32 crc kubenswrapper[4967]: I1011 04:07:32.181453 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=4.181435357 podStartE2EDuration="4.181435357s" podCreationTimestamp="2025-10-11 04:07:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:07:32.176214007 +0000 UTC m=+980.139422950" watchObservedRunningTime="2025-10-11 04:07:32.181435357 +0000 UTC m=+980.144644300" Oct 11 04:07:32 crc kubenswrapper[4967]: I1011 04:07:32.183022 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3ddabca8-7c71-431e-b3f6-987d21900b0d" containerName="glance-log" containerID="cri-o://7fc00872eb20d715869841e872d53e5203e2c809679a6cf3edc55987d07c4441" gracePeriod=30 Oct 11 04:07:32 crc kubenswrapper[4967]: I1011 04:07:32.183229 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="3ddabca8-7c71-431e-b3f6-987d21900b0d" containerName="glance-httpd" containerID="cri-o://3c1123635771bf7e931bd9889850148e163ce9c271fedcbabe1846b6549fdd39" gracePeriod=30 Oct 11 04:07:32 crc kubenswrapper[4967]: I1011 04:07:32.183300 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3ddabca8-7c71-431e-b3f6-987d21900b0d","Type":"ContainerStarted","Data":"3c1123635771bf7e931bd9889850148e163ce9c271fedcbabe1846b6549fdd39"} Oct 11 04:07:32 crc kubenswrapper[4967]: I1011 04:07:32.210062 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=4.210037226 podStartE2EDuration="4.210037226s" podCreationTimestamp="2025-10-11 04:07:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:07:32.204583751 +0000 UTC m=+980.167792694" watchObservedRunningTime="2025-10-11 04:07:32.210037226 +0000 UTC m=+980.173246159" Oct 11 04:07:32 crc kubenswrapper[4967]: I1011 04:07:32.647436 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-b879-account-create-6v2r9"] Oct 11 04:07:32 crc kubenswrapper[4967]: E1011 04:07:32.647847 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="c70d7931-0ef5-44dc-88dd-b0f7b5ef3afd" containerName="mariadb-database-create" Oct 11 04:07:32 crc kubenswrapper[4967]: I1011 04:07:32.647867 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="c70d7931-0ef5-44dc-88dd-b0f7b5ef3afd" containerName="mariadb-database-create" Oct 11 04:07:32 crc kubenswrapper[4967]: E1011 04:07:32.647886 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="43f1cabd-1a2d-4996-98e4-d90f58007fab" containerName="init" Oct 11 04:07:32 crc kubenswrapper[4967]: I1011 04:07:32.647896 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="43f1cabd-1a2d-4996-98e4-d90f58007fab" containerName="init" Oct 11 04:07:32 crc kubenswrapper[4967]: I1011 04:07:32.648122 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="c70d7931-0ef5-44dc-88dd-b0f7b5ef3afd" containerName="mariadb-database-create" Oct 11 04:07:32 crc kubenswrapper[4967]: I1011 04:07:32.648148 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="43f1cabd-1a2d-4996-98e4-d90f58007fab" containerName="init" Oct 11 04:07:32 crc kubenswrapper[4967]: I1011 04:07:32.648797 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-b879-account-create-6v2r9" Oct 11 04:07:32 crc kubenswrapper[4967]: I1011 04:07:32.650494 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-db-secret" Oct 11 04:07:32 crc kubenswrapper[4967]: I1011 04:07:32.655203 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-b879-account-create-6v2r9"] Oct 11 04:07:32 crc kubenswrapper[4967]: I1011 04:07:32.746138 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-5b20-account-create-lb22f"] Oct 11 04:07:32 crc kubenswrapper[4967]: I1011 04:07:32.747541 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5b20-account-create-lb22f" Oct 11 04:07:32 crc kubenswrapper[4967]: I1011 04:07:32.750622 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-db-secret" Oct 11 04:07:32 crc kubenswrapper[4967]: I1011 04:07:32.763219 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slch9\" (UniqueName: \"kubernetes.io/projected/92ccb4c4-e0b5-4592-bdfc-9909ee48b8b1-kube-api-access-slch9\") pod \"cinder-b879-account-create-6v2r9\" (UID: \"92ccb4c4-e0b5-4592-bdfc-9909ee48b8b1\") " pod="openstack/cinder-b879-account-create-6v2r9" Oct 11 04:07:32 crc kubenswrapper[4967]: I1011 04:07:32.767112 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-5b20-account-create-lb22f"] Oct 11 04:07:32 crc kubenswrapper[4967]: I1011 04:07:32.864770 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slch9\" (UniqueName: \"kubernetes.io/projected/92ccb4c4-e0b5-4592-bdfc-9909ee48b8b1-kube-api-access-slch9\") pod \"cinder-b879-account-create-6v2r9\" (UID: \"92ccb4c4-e0b5-4592-bdfc-9909ee48b8b1\") " pod="openstack/cinder-b879-account-create-6v2r9" Oct 11 04:07:32 crc kubenswrapper[4967]: I1011 04:07:32.864892 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lnv6d\" (UniqueName: \"kubernetes.io/projected/2743338e-1478-4402-b29f-d4b86722d241-kube-api-access-lnv6d\") pod \"barbican-5b20-account-create-lb22f\" (UID: \"2743338e-1478-4402-b29f-d4b86722d241\") " pod="openstack/barbican-5b20-account-create-lb22f" Oct 11 04:07:32 crc kubenswrapper[4967]: I1011 04:07:32.885326 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slch9\" (UniqueName: \"kubernetes.io/projected/92ccb4c4-e0b5-4592-bdfc-9909ee48b8b1-kube-api-access-slch9\") pod \"cinder-b879-account-create-6v2r9\" (UID: \"92ccb4c4-e0b5-4592-bdfc-9909ee48b8b1\") " pod="openstack/cinder-b879-account-create-6v2r9" Oct 11 04:07:32 crc kubenswrapper[4967]: I1011 04:07:32.967377 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lnv6d\" (UniqueName: \"kubernetes.io/projected/2743338e-1478-4402-b29f-d4b86722d241-kube-api-access-lnv6d\") pod \"barbican-5b20-account-create-lb22f\" (UID: \"2743338e-1478-4402-b29f-d4b86722d241\") " pod="openstack/barbican-5b20-account-create-lb22f" Oct 11 04:07:32 crc kubenswrapper[4967]: I1011 04:07:32.984560 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lnv6d\" (UniqueName: \"kubernetes.io/projected/2743338e-1478-4402-b29f-d4b86722d241-kube-api-access-lnv6d\") pod \"barbican-5b20-account-create-lb22f\" (UID: \"2743338e-1478-4402-b29f-d4b86722d241\") " pod="openstack/barbican-5b20-account-create-lb22f" Oct 11 04:07:33 crc kubenswrapper[4967]: I1011 04:07:33.023457 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-b879-account-create-6v2r9" Oct 11 04:07:33 crc kubenswrapper[4967]: I1011 04:07:33.045153 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-317e-account-create-l88bq"] Oct 11 04:07:33 crc kubenswrapper[4967]: I1011 04:07:33.047110 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-317e-account-create-l88bq" Oct 11 04:07:33 crc kubenswrapper[4967]: I1011 04:07:33.052294 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-db-secret" Oct 11 04:07:33 crc kubenswrapper[4967]: I1011 04:07:33.053325 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-317e-account-create-l88bq"] Oct 11 04:07:33 crc kubenswrapper[4967]: I1011 04:07:33.081704 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5b20-account-create-lb22f" Oct 11 04:07:33 crc kubenswrapper[4967]: I1011 04:07:33.170487 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-c5z8c\" (UniqueName: \"kubernetes.io/projected/2930d6b7-0aad-4123-b8bd-4dad4b8768fe-kube-api-access-c5z8c\") pod \"neutron-317e-account-create-l88bq\" (UID: \"2930d6b7-0aad-4123-b8bd-4dad4b8768fe\") " pod="openstack/neutron-317e-account-create-l88bq" Oct 11 04:07:33 crc kubenswrapper[4967]: I1011 04:07:33.192668 4967 generic.go:334] "Generic (PLEG): container finished" podID="b92605b2-50a6-41bc-8a34-8783cc9fc7c9" containerID="dea40148c9a1a67da3b4bd3d86f7758d94d9b0d3207421731cb4c1e6dbc02697" exitCode=0 Oct 11 04:07:33 crc kubenswrapper[4967]: I1011 04:07:33.192704 4967 generic.go:334] "Generic (PLEG): container finished" podID="b92605b2-50a6-41bc-8a34-8783cc9fc7c9" containerID="e06f2ac89aa25dc24dc2dab0cb337b7fe9eafd14405c4942ed856a86f494bb9d" exitCode=143 Oct 11 04:07:33 crc kubenswrapper[4967]: I1011 04:07:33.192750 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b92605b2-50a6-41bc-8a34-8783cc9fc7c9","Type":"ContainerDied","Data":"dea40148c9a1a67da3b4bd3d86f7758d94d9b0d3207421731cb4c1e6dbc02697"} Oct 11 04:07:33 crc kubenswrapper[4967]: I1011 04:07:33.192791 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b92605b2-50a6-41bc-8a34-8783cc9fc7c9","Type":"ContainerDied","Data":"e06f2ac89aa25dc24dc2dab0cb337b7fe9eafd14405c4942ed856a86f494bb9d"} Oct 11 04:07:33 crc kubenswrapper[4967]: I1011 04:07:33.195106 4967 generic.go:334] "Generic (PLEG): container finished" podID="31640236-7dcb-4bd8-851b-0b5870950b02" containerID="8d452e5bcdee06ff27ad299bd16449ebf91079000282280714301c18395e32a6" exitCode=0 Oct 11 04:07:33 crc kubenswrapper[4967]: I1011 04:07:33.195162 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jmt45" event={"ID":"31640236-7dcb-4bd8-851b-0b5870950b02","Type":"ContainerDied","Data":"8d452e5bcdee06ff27ad299bd16449ebf91079000282280714301c18395e32a6"} Oct 11 04:07:33 crc kubenswrapper[4967]: I1011 04:07:33.198441 4967 generic.go:334] "Generic (PLEG): container finished" podID="3ddabca8-7c71-431e-b3f6-987d21900b0d" containerID="3c1123635771bf7e931bd9889850148e163ce9c271fedcbabe1846b6549fdd39" exitCode=0 Oct 11 04:07:33 crc kubenswrapper[4967]: I1011 04:07:33.198467 4967 generic.go:334] "Generic (PLEG): container finished" podID="3ddabca8-7c71-431e-b3f6-987d21900b0d" containerID="7fc00872eb20d715869841e872d53e5203e2c809679a6cf3edc55987d07c4441" exitCode=143 Oct 11 04:07:33 crc kubenswrapper[4967]: I1011 04:07:33.198487 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3ddabca8-7c71-431e-b3f6-987d21900b0d","Type":"ContainerDied","Data":"3c1123635771bf7e931bd9889850148e163ce9c271fedcbabe1846b6549fdd39"} Oct 11 04:07:33 crc kubenswrapper[4967]: I1011 04:07:33.198507 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3ddabca8-7c71-431e-b3f6-987d21900b0d","Type":"ContainerDied","Data":"7fc00872eb20d715869841e872d53e5203e2c809679a6cf3edc55987d07c4441"} Oct 11 04:07:33 crc kubenswrapper[4967]: I1011 04:07:33.271810 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-c5z8c\" (UniqueName: \"kubernetes.io/projected/2930d6b7-0aad-4123-b8bd-4dad4b8768fe-kube-api-access-c5z8c\") pod \"neutron-317e-account-create-l88bq\" (UID: \"2930d6b7-0aad-4123-b8bd-4dad4b8768fe\") " pod="openstack/neutron-317e-account-create-l88bq" Oct 11 04:07:33 crc kubenswrapper[4967]: I1011 04:07:33.291383 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-c5z8c\" (UniqueName: \"kubernetes.io/projected/2930d6b7-0aad-4123-b8bd-4dad4b8768fe-kube-api-access-c5z8c\") pod \"neutron-317e-account-create-l88bq\" (UID: \"2930d6b7-0aad-4123-b8bd-4dad4b8768fe\") " pod="openstack/neutron-317e-account-create-l88bq" Oct 11 04:07:33 crc kubenswrapper[4967]: I1011 04:07:33.370535 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-317e-account-create-l88bq" Oct 11 04:07:36 crc kubenswrapper[4967]: I1011 04:07:36.231793 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-jmt45" event={"ID":"31640236-7dcb-4bd8-851b-0b5870950b02","Type":"ContainerDied","Data":"29b3adc73ac1f1ac2e7537e5f0d514f3884577e5da05b844bde9d925e62fa6bc"} Oct 11 04:07:36 crc kubenswrapper[4967]: I1011 04:07:36.232489 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="29b3adc73ac1f1ac2e7537e5f0d514f3884577e5da05b844bde9d925e62fa6bc" Oct 11 04:07:36 crc kubenswrapper[4967]: I1011 04:07:36.236723 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"b92605b2-50a6-41bc-8a34-8783cc9fc7c9","Type":"ContainerDied","Data":"a8c9b656af8d8321b27ecdb510d6f332ff15bd517bb4746fab1589a3736a3fdd"} Oct 11 04:07:36 crc kubenswrapper[4967]: I1011 04:07:36.236768 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a8c9b656af8d8321b27ecdb510d6f332ff15bd517bb4746fab1589a3736a3fdd" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.264639 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jmt45" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.286357 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.428694 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\" (UID: \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\") " Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.429863 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-combined-ca-bundle\") pod \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\" (UID: \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\") " Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.429906 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31640236-7dcb-4bd8-851b-0b5870950b02-combined-ca-bundle\") pod \"31640236-7dcb-4bd8-851b-0b5870950b02\" (UID: \"31640236-7dcb-4bd8-851b-0b5870950b02\") " Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.429933 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-scripts\") pod \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\" (UID: \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\") " Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.430287 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31640236-7dcb-4bd8-851b-0b5870950b02-config-data\") pod \"31640236-7dcb-4bd8-851b-0b5870950b02\" (UID: \"31640236-7dcb-4bd8-851b-0b5870950b02\") " Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.430304 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-config-data\") pod \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\" (UID: \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\") " Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.430349 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31640236-7dcb-4bd8-851b-0b5870950b02-scripts\") pod \"31640236-7dcb-4bd8-851b-0b5870950b02\" (UID: \"31640236-7dcb-4bd8-851b-0b5870950b02\") " Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.430384 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-logs\") pod \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\" (UID: \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\") " Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.430436 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-httpd-run\") pod \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\" (UID: \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\") " Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.430469 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ss5pf\" (UniqueName: \"kubernetes.io/projected/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-kube-api-access-ss5pf\") pod \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\" (UID: \"b92605b2-50a6-41bc-8a34-8783cc9fc7c9\") " Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.430791 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/31640236-7dcb-4bd8-851b-0b5870950b02-fernet-keys\") pod \"31640236-7dcb-4bd8-851b-0b5870950b02\" (UID: \"31640236-7dcb-4bd8-851b-0b5870950b02\") " Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.430807 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2nzdj\" (UniqueName: \"kubernetes.io/projected/31640236-7dcb-4bd8-851b-0b5870950b02-kube-api-access-2nzdj\") pod \"31640236-7dcb-4bd8-851b-0b5870950b02\" (UID: \"31640236-7dcb-4bd8-851b-0b5870950b02\") " Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.430834 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/31640236-7dcb-4bd8-851b-0b5870950b02-credential-keys\") pod \"31640236-7dcb-4bd8-851b-0b5870950b02\" (UID: \"31640236-7dcb-4bd8-851b-0b5870950b02\") " Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.431224 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-logs" (OuterVolumeSpecName: "logs") pod "b92605b2-50a6-41bc-8a34-8783cc9fc7c9" (UID: "b92605b2-50a6-41bc-8a34-8783cc9fc7c9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.431552 4967 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-logs\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.431852 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "b92605b2-50a6-41bc-8a34-8783cc9fc7c9" (UID: "b92605b2-50a6-41bc-8a34-8783cc9fc7c9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.436673 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31640236-7dcb-4bd8-851b-0b5870950b02-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "31640236-7dcb-4bd8-851b-0b5870950b02" (UID: "31640236-7dcb-4bd8-851b-0b5870950b02"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.436746 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "b92605b2-50a6-41bc-8a34-8783cc9fc7c9" (UID: "b92605b2-50a6-41bc-8a34-8783cc9fc7c9"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.436750 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-scripts" (OuterVolumeSpecName: "scripts") pod "b92605b2-50a6-41bc-8a34-8783cc9fc7c9" (UID: "b92605b2-50a6-41bc-8a34-8783cc9fc7c9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.437450 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31640236-7dcb-4bd8-851b-0b5870950b02-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "31640236-7dcb-4bd8-851b-0b5870950b02" (UID: "31640236-7dcb-4bd8-851b-0b5870950b02"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.437507 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31640236-7dcb-4bd8-851b-0b5870950b02-kube-api-access-2nzdj" (OuterVolumeSpecName: "kube-api-access-2nzdj") pod "31640236-7dcb-4bd8-851b-0b5870950b02" (UID: "31640236-7dcb-4bd8-851b-0b5870950b02"). InnerVolumeSpecName "kube-api-access-2nzdj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.440169 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-kube-api-access-ss5pf" (OuterVolumeSpecName: "kube-api-access-ss5pf") pod "b92605b2-50a6-41bc-8a34-8783cc9fc7c9" (UID: "b92605b2-50a6-41bc-8a34-8783cc9fc7c9"). InnerVolumeSpecName "kube-api-access-ss5pf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.444315 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31640236-7dcb-4bd8-851b-0b5870950b02-scripts" (OuterVolumeSpecName: "scripts") pod "31640236-7dcb-4bd8-851b-0b5870950b02" (UID: "31640236-7dcb-4bd8-851b-0b5870950b02"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.464918 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "b92605b2-50a6-41bc-8a34-8783cc9fc7c9" (UID: "b92605b2-50a6-41bc-8a34-8783cc9fc7c9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.466985 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31640236-7dcb-4bd8-851b-0b5870950b02-config-data" (OuterVolumeSpecName: "config-data") pod "31640236-7dcb-4bd8-851b-0b5870950b02" (UID: "31640236-7dcb-4bd8-851b-0b5870950b02"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.480622 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31640236-7dcb-4bd8-851b-0b5870950b02-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "31640236-7dcb-4bd8-851b-0b5870950b02" (UID: "31640236-7dcb-4bd8-851b-0b5870950b02"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.486306 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-config-data" (OuterVolumeSpecName: "config-data") pod "b92605b2-50a6-41bc-8a34-8783cc9fc7c9" (UID: "b92605b2-50a6-41bc-8a34-8783cc9fc7c9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.545283 4967 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/31640236-7dcb-4bd8-851b-0b5870950b02-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.545319 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2nzdj\" (UniqueName: \"kubernetes.io/projected/31640236-7dcb-4bd8-851b-0b5870950b02-kube-api-access-2nzdj\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.545332 4967 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/31640236-7dcb-4bd8-851b-0b5870950b02-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.545362 4967 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.545375 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.545386 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/31640236-7dcb-4bd8-851b-0b5870950b02-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.545397 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.545408 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/31640236-7dcb-4bd8-851b-0b5870950b02-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.545418 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.545427 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/31640236-7dcb-4bd8-851b-0b5870950b02-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.545436 4967 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.545446 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ss5pf\" (UniqueName: \"kubernetes.io/projected/b92605b2-50a6-41bc-8a34-8783cc9fc7c9-kube-api-access-ss5pf\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.567775 4967 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:36.647090 4967 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.245045 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zf6xs" event={"ID":"05059c3c-e29a-4bfa-a0d4-0d56b3d85af9","Type":"ContainerStarted","Data":"abdc71002731f1f79b9c3e24a5bce139cf8e8ef69665b9d7a89453b7eda936ce"} Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.252102 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"3ddabca8-7c71-431e-b3f6-987d21900b0d","Type":"ContainerDied","Data":"e6b952693f5a79981f71f9a4424a7e1b7b1443dff3d37e49cd3bee4a22ae1e68"} Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.252139 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e6b952693f5a79981f71f9a4424a7e1b7b1443dff3d37e49cd3bee4a22ae1e68" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.253356 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-jmt45" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.253913 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb56ccd5-8cd3-4da0-b11f-8600af8abf86","Type":"ContainerStarted","Data":"26117f8c594b515e35c7c2108fef08a9638b8113dcdf4be3945b880c53864c0d"} Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.253955 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.278496 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-db-sync-zf6xs" podStartSLOduration=2.303228315 podStartE2EDuration="9.278476615s" podCreationTimestamp="2025-10-11 04:07:28 +0000 UTC" firstStartedPulling="2025-10-11 04:07:29.787028782 +0000 UTC m=+977.750237715" lastFinishedPulling="2025-10-11 04:07:36.762277082 +0000 UTC m=+984.725486015" observedRunningTime="2025-10-11 04:07:37.277383391 +0000 UTC m=+985.240592324" watchObservedRunningTime="2025-10-11 04:07:37.278476615 +0000 UTC m=+985.241685548" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.300189 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.337947 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.347449 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.357163 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 11 04:07:37 crc kubenswrapper[4967]: E1011 04:07:37.357622 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ddabca8-7c71-431e-b3f6-987d21900b0d" containerName="glance-httpd" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.357645 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ddabca8-7c71-431e-b3f6-987d21900b0d" containerName="glance-httpd" Oct 11 04:07:37 crc kubenswrapper[4967]: E1011 04:07:37.357669 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31640236-7dcb-4bd8-851b-0b5870950b02" containerName="keystone-bootstrap" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.357679 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="31640236-7dcb-4bd8-851b-0b5870950b02" containerName="keystone-bootstrap" Oct 11 04:07:37 crc kubenswrapper[4967]: E1011 04:07:37.357709 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b92605b2-50a6-41bc-8a34-8783cc9fc7c9" containerName="glance-log" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.357719 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="b92605b2-50a6-41bc-8a34-8783cc9fc7c9" containerName="glance-log" Oct 11 04:07:37 crc kubenswrapper[4967]: E1011 04:07:37.357732 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ddabca8-7c71-431e-b3f6-987d21900b0d" containerName="glance-log" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.357738 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ddabca8-7c71-431e-b3f6-987d21900b0d" containerName="glance-log" Oct 11 04:07:37 crc kubenswrapper[4967]: E1011 04:07:37.357757 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b92605b2-50a6-41bc-8a34-8783cc9fc7c9" containerName="glance-httpd" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.357765 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="b92605b2-50a6-41bc-8a34-8783cc9fc7c9" containerName="glance-httpd" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.357970 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ddabca8-7c71-431e-b3f6-987d21900b0d" containerName="glance-httpd" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.357993 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="b92605b2-50a6-41bc-8a34-8783cc9fc7c9" containerName="glance-httpd" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.358011 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="31640236-7dcb-4bd8-851b-0b5870950b02" containerName="keystone-bootstrap" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.358021 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ddabca8-7c71-431e-b3f6-987d21900b0d" containerName="glance-log" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.358031 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="b92605b2-50a6-41bc-8a34-8783cc9fc7c9" containerName="glance-log" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.358910 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.372782 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.385575 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.441216 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-317e-account-create-l88bq"] Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.449830 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-jmt45"] Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.462148 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-jmt45"] Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.476127 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-5b20-account-create-lb22f"] Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.481563 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ddabca8-7c71-431e-b3f6-987d21900b0d-config-data\") pod \"3ddabca8-7c71-431e-b3f6-987d21900b0d\" (UID: \"3ddabca8-7c71-431e-b3f6-987d21900b0d\") " Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.481615 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3ddabca8-7c71-431e-b3f6-987d21900b0d-httpd-run\") pod \"3ddabca8-7c71-431e-b3f6-987d21900b0d\" (UID: \"3ddabca8-7c71-431e-b3f6-987d21900b0d\") " Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.481739 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ddabca8-7c71-431e-b3f6-987d21900b0d-scripts\") pod \"3ddabca8-7c71-431e-b3f6-987d21900b0d\" (UID: \"3ddabca8-7c71-431e-b3f6-987d21900b0d\") " Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.481788 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zq7mz\" (UniqueName: \"kubernetes.io/projected/3ddabca8-7c71-431e-b3f6-987d21900b0d-kube-api-access-zq7mz\") pod \"3ddabca8-7c71-431e-b3f6-987d21900b0d\" (UID: \"3ddabca8-7c71-431e-b3f6-987d21900b0d\") " Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.481818 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ddabca8-7c71-431e-b3f6-987d21900b0d-logs\") pod \"3ddabca8-7c71-431e-b3f6-987d21900b0d\" (UID: \"3ddabca8-7c71-431e-b3f6-987d21900b0d\") " Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.481882 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ddabca8-7c71-431e-b3f6-987d21900b0d-combined-ca-bundle\") pod \"3ddabca8-7c71-431e-b3f6-987d21900b0d\" (UID: \"3ddabca8-7c71-431e-b3f6-987d21900b0d\") " Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.481992 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"3ddabca8-7c71-431e-b3f6-987d21900b0d\" (UID: \"3ddabca8-7c71-431e-b3f6-987d21900b0d\") " Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.482207 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5772c259-389a-4faf-8307-f14eff408981-logs\") pod \"glance-default-internal-api-0\" (UID: \"5772c259-389a-4faf-8307-f14eff408981\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.482248 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"5772c259-389a-4faf-8307-f14eff408981\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.482306 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5772c259-389a-4faf-8307-f14eff408981-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5772c259-389a-4faf-8307-f14eff408981\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.482327 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5772c259-389a-4faf-8307-f14eff408981-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5772c259-389a-4faf-8307-f14eff408981\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.482394 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5772c259-389a-4faf-8307-f14eff408981-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5772c259-389a-4faf-8307-f14eff408981\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.482416 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6cxvs\" (UniqueName: \"kubernetes.io/projected/5772c259-389a-4faf-8307-f14eff408981-kube-api-access-6cxvs\") pod \"glance-default-internal-api-0\" (UID: \"5772c259-389a-4faf-8307-f14eff408981\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.482440 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5772c259-389a-4faf-8307-f14eff408981-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5772c259-389a-4faf-8307-f14eff408981\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.485145 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ddabca8-7c71-431e-b3f6-987d21900b0d-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "3ddabca8-7c71-431e-b3f6-987d21900b0d" (UID: "3ddabca8-7c71-431e-b3f6-987d21900b0d"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.486108 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3ddabca8-7c71-431e-b3f6-987d21900b0d-logs" (OuterVolumeSpecName: "logs") pod "3ddabca8-7c71-431e-b3f6-987d21900b0d" (UID: "3ddabca8-7c71-431e-b3f6-987d21900b0d"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.489164 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ddabca8-7c71-431e-b3f6-987d21900b0d-scripts" (OuterVolumeSpecName: "scripts") pod "3ddabca8-7c71-431e-b3f6-987d21900b0d" (UID: "3ddabca8-7c71-431e-b3f6-987d21900b0d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.489716 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ddabca8-7c71-431e-b3f6-987d21900b0d-kube-api-access-zq7mz" (OuterVolumeSpecName: "kube-api-access-zq7mz") pod "3ddabca8-7c71-431e-b3f6-987d21900b0d" (UID: "3ddabca8-7c71-431e-b3f6-987d21900b0d"). InnerVolumeSpecName "kube-api-access-zq7mz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.491640 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-b879-account-create-6v2r9"] Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.491992 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "3ddabca8-7c71-431e-b3f6-987d21900b0d" (UID: "3ddabca8-7c71-431e-b3f6-987d21900b0d"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.529886 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ddabca8-7c71-431e-b3f6-987d21900b0d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3ddabca8-7c71-431e-b3f6-987d21900b0d" (UID: "3ddabca8-7c71-431e-b3f6-987d21900b0d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.547507 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-bootstrap-5dtlx"] Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.550049 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5dtlx" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.560153 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-5dtlx"] Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.563297 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.571662 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.574003 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.574369 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-k7kxn" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.585030 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"5772c259-389a-4faf-8307-f14eff408981\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.585142 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5772c259-389a-4faf-8307-f14eff408981-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5772c259-389a-4faf-8307-f14eff408981\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.585167 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5772c259-389a-4faf-8307-f14eff408981-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5772c259-389a-4faf-8307-f14eff408981\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.585229 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5772c259-389a-4faf-8307-f14eff408981-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5772c259-389a-4faf-8307-f14eff408981\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.585255 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6cxvs\" (UniqueName: \"kubernetes.io/projected/5772c259-389a-4faf-8307-f14eff408981-kube-api-access-6cxvs\") pod \"glance-default-internal-api-0\" (UID: \"5772c259-389a-4faf-8307-f14eff408981\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.585284 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5772c259-389a-4faf-8307-f14eff408981-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5772c259-389a-4faf-8307-f14eff408981\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.586456 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5772c259-389a-4faf-8307-f14eff408981-logs\") pod \"glance-default-internal-api-0\" (UID: \"5772c259-389a-4faf-8307-f14eff408981\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.586544 4967 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"5772c259-389a-4faf-8307-f14eff408981\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.586592 4967 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/3ddabca8-7c71-431e-b3f6-987d21900b0d-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.587296 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3ddabca8-7c71-431e-b3f6-987d21900b0d-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.587313 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zq7mz\" (UniqueName: \"kubernetes.io/projected/3ddabca8-7c71-431e-b3f6-987d21900b0d-kube-api-access-zq7mz\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.587324 4967 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/3ddabca8-7c71-431e-b3f6-987d21900b0d-logs\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.587342 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3ddabca8-7c71-431e-b3f6-987d21900b0d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.587370 4967 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.588393 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5772c259-389a-4faf-8307-f14eff408981-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"5772c259-389a-4faf-8307-f14eff408981\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.588496 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5772c259-389a-4faf-8307-f14eff408981-logs\") pod \"glance-default-internal-api-0\" (UID: \"5772c259-389a-4faf-8307-f14eff408981\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.593674 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5772c259-389a-4faf-8307-f14eff408981-scripts\") pod \"glance-default-internal-api-0\" (UID: \"5772c259-389a-4faf-8307-f14eff408981\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.606283 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5772c259-389a-4faf-8307-f14eff408981-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"5772c259-389a-4faf-8307-f14eff408981\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.606757 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5772c259-389a-4faf-8307-f14eff408981-config-data\") pod \"glance-default-internal-api-0\" (UID: \"5772c259-389a-4faf-8307-f14eff408981\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.606890 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6cxvs\" (UniqueName: \"kubernetes.io/projected/5772c259-389a-4faf-8307-f14eff408981-kube-api-access-6cxvs\") pod \"glance-default-internal-api-0\" (UID: \"5772c259-389a-4faf-8307-f14eff408981\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.637099 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ddabca8-7c71-431e-b3f6-987d21900b0d-config-data" (OuterVolumeSpecName: "config-data") pod "3ddabca8-7c71-431e-b3f6-987d21900b0d" (UID: "3ddabca8-7c71-431e-b3f6-987d21900b0d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.652551 4967 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.655336 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"5772c259-389a-4faf-8307-f14eff408981\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.659555 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.660199 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.691781 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-fernet-keys\") pod \"keystone-bootstrap-5dtlx\" (UID: \"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3\") " pod="openstack/keystone-bootstrap-5dtlx" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.691840 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-scripts\") pod \"keystone-bootstrap-5dtlx\" (UID: \"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3\") " pod="openstack/keystone-bootstrap-5dtlx" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.691938 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-credential-keys\") pod \"keystone-bootstrap-5dtlx\" (UID: \"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3\") " pod="openstack/keystone-bootstrap-5dtlx" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.691993 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bp58w\" (UniqueName: \"kubernetes.io/projected/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-kube-api-access-bp58w\") pod \"keystone-bootstrap-5dtlx\" (UID: \"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3\") " pod="openstack/keystone-bootstrap-5dtlx" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.692062 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-config-data\") pod \"keystone-bootstrap-5dtlx\" (UID: \"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3\") " pod="openstack/keystone-bootstrap-5dtlx" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.692115 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-combined-ca-bundle\") pod \"keystone-bootstrap-5dtlx\" (UID: \"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3\") " pod="openstack/keystone-bootstrap-5dtlx" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.692203 4967 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.692226 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3ddabca8-7c71-431e-b3f6-987d21900b0d-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.793273 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-config-data\") pod \"keystone-bootstrap-5dtlx\" (UID: \"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3\") " pod="openstack/keystone-bootstrap-5dtlx" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.793632 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-combined-ca-bundle\") pod \"keystone-bootstrap-5dtlx\" (UID: \"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3\") " pod="openstack/keystone-bootstrap-5dtlx" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.793674 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-fernet-keys\") pod \"keystone-bootstrap-5dtlx\" (UID: \"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3\") " pod="openstack/keystone-bootstrap-5dtlx" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.793701 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-scripts\") pod \"keystone-bootstrap-5dtlx\" (UID: \"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3\") " pod="openstack/keystone-bootstrap-5dtlx" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.793755 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-credential-keys\") pod \"keystone-bootstrap-5dtlx\" (UID: \"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3\") " pod="openstack/keystone-bootstrap-5dtlx" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.793796 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bp58w\" (UniqueName: \"kubernetes.io/projected/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-kube-api-access-bp58w\") pod \"keystone-bootstrap-5dtlx\" (UID: \"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3\") " pod="openstack/keystone-bootstrap-5dtlx" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.802186 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-credential-keys\") pod \"keystone-bootstrap-5dtlx\" (UID: \"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3\") " pod="openstack/keystone-bootstrap-5dtlx" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.806382 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-scripts\") pod \"keystone-bootstrap-5dtlx\" (UID: \"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3\") " pod="openstack/keystone-bootstrap-5dtlx" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.813694 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bp58w\" (UniqueName: \"kubernetes.io/projected/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-kube-api-access-bp58w\") pod \"keystone-bootstrap-5dtlx\" (UID: \"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3\") " pod="openstack/keystone-bootstrap-5dtlx" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.815653 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-fernet-keys\") pod \"keystone-bootstrap-5dtlx\" (UID: \"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3\") " pod="openstack/keystone-bootstrap-5dtlx" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.816216 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-config-data\") pod \"keystone-bootstrap-5dtlx\" (UID: \"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3\") " pod="openstack/keystone-bootstrap-5dtlx" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.823199 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-combined-ca-bundle\") pod \"keystone-bootstrap-5dtlx\" (UID: \"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3\") " pod="openstack/keystone-bootstrap-5dtlx" Oct 11 04:07:37 crc kubenswrapper[4967]: I1011 04:07:37.895280 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5dtlx" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.133654 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-bootstrap-5dtlx"] Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.192147 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.272605 4967 generic.go:334] "Generic (PLEG): container finished" podID="92ccb4c4-e0b5-4592-bdfc-9909ee48b8b1" containerID="d0751b3ff58d0130b73e6aa500956566e743d151ef828a30c0b59e0ae23c9a49" exitCode=0 Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.272661 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-b879-account-create-6v2r9" event={"ID":"92ccb4c4-e0b5-4592-bdfc-9909ee48b8b1","Type":"ContainerDied","Data":"d0751b3ff58d0130b73e6aa500956566e743d151ef828a30c0b59e0ae23c9a49"} Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.272685 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-b879-account-create-6v2r9" event={"ID":"92ccb4c4-e0b5-4592-bdfc-9909ee48b8b1","Type":"ContainerStarted","Data":"c85053d0d5508ed96808221409a03c305dd73770f11fe1bbfe75ab4b7c89324d"} Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.283520 4967 generic.go:334] "Generic (PLEG): container finished" podID="2930d6b7-0aad-4123-b8bd-4dad4b8768fe" containerID="31a36b28144237ec04957225c745f8fc483b4ed7206adce1567814f143f87402" exitCode=0 Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.283620 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-317e-account-create-l88bq" event={"ID":"2930d6b7-0aad-4123-b8bd-4dad4b8768fe","Type":"ContainerDied","Data":"31a36b28144237ec04957225c745f8fc483b4ed7206adce1567814f143f87402"} Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.283647 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-317e-account-create-l88bq" event={"ID":"2930d6b7-0aad-4123-b8bd-4dad4b8768fe","Type":"ContainerStarted","Data":"ed6ebe5c5273dec147dc0d9f8cdbdade6490dc57c902315056fea30bdc3edd32"} Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.295110 4967 generic.go:334] "Generic (PLEG): container finished" podID="2743338e-1478-4402-b29f-d4b86722d241" containerID="3a312f975d52e911748a61f193ab7ac5430ad2106363aacfd23680c68c6d79a9" exitCode=0 Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.295187 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5b20-account-create-lb22f" event={"ID":"2743338e-1478-4402-b29f-d4b86722d241","Type":"ContainerDied","Data":"3a312f975d52e911748a61f193ab7ac5430ad2106363aacfd23680c68c6d79a9"} Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.295216 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.295231 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5b20-account-create-lb22f" event={"ID":"2743338e-1478-4402-b29f-d4b86722d241","Type":"ContainerStarted","Data":"9af6783eb9a6a545ddbd706d12233acccfe85b00f45bceda347ec168058969ae"} Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.357367 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.365168 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.374390 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.377513 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.383222 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.383506 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.389147 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.418326 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-8389-account-create-pvbnb"] Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.421251 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-8389-account-create-pvbnb" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.424212 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-db-secret" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.427117 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-8389-account-create-pvbnb"] Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.506317 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-config-data\") pod \"glance-default-external-api-0\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.506531 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hgz4k\" (UniqueName: \"kubernetes.io/projected/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-kube-api-access-hgz4k\") pod \"glance-default-external-api-0\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.506613 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.506673 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-logs\") pod \"glance-default-external-api-0\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.506860 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-scripts\") pod \"glance-default-external-api-0\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.506979 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-272zg\" (UniqueName: \"kubernetes.io/projected/4691fb42-e6bb-4e1b-a4c2-1f65e1292d4d-kube-api-access-272zg\") pod \"ironic-8389-account-create-pvbnb\" (UID: \"4691fb42-e6bb-4e1b-a4c2-1f65e1292d4d\") " pod="openstack/ironic-8389-account-create-pvbnb" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.507049 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.507135 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.507189 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.608958 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-config-data\") pod \"glance-default-external-api-0\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.609001 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hgz4k\" (UniqueName: \"kubernetes.io/projected/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-kube-api-access-hgz4k\") pod \"glance-default-external-api-0\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.609035 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.609052 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-logs\") pod \"glance-default-external-api-0\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.609136 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-scripts\") pod \"glance-default-external-api-0\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.609173 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-272zg\" (UniqueName: \"kubernetes.io/projected/4691fb42-e6bb-4e1b-a4c2-1f65e1292d4d-kube-api-access-272zg\") pod \"ironic-8389-account-create-pvbnb\" (UID: \"4691fb42-e6bb-4e1b-a4c2-1f65e1292d4d\") " pod="openstack/ironic-8389-account-create-pvbnb" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.609203 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.609230 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.609255 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.609671 4967 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.610055 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.610345 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-logs\") pod \"glance-default-external-api-0\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.618680 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.623308 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.625307 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-config-data\") pod \"glance-default-external-api-0\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.625312 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-scripts\") pod \"glance-default-external-api-0\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.645591 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hgz4k\" (UniqueName: \"kubernetes.io/projected/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-kube-api-access-hgz4k\") pod \"glance-default-external-api-0\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.648474 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") " pod="openstack/glance-default-external-api-0" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.649552 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-272zg\" (UniqueName: \"kubernetes.io/projected/4691fb42-e6bb-4e1b-a4c2-1f65e1292d4d-kube-api-access-272zg\") pod \"ironic-8389-account-create-pvbnb\" (UID: \"4691fb42-e6bb-4e1b-a4c2-1f65e1292d4d\") " pod="openstack/ironic-8389-account-create-pvbnb" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.707159 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.742386 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-8389-account-create-pvbnb" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.833306 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31640236-7dcb-4bd8-851b-0b5870950b02" path="/var/lib/kubelet/pods/31640236-7dcb-4bd8-851b-0b5870950b02/volumes" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.834061 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ddabca8-7c71-431e-b3f6-987d21900b0d" path="/var/lib/kubelet/pods/3ddabca8-7c71-431e-b3f6-987d21900b0d/volumes" Oct 11 04:07:38 crc kubenswrapper[4967]: I1011 04:07:38.835206 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b92605b2-50a6-41bc-8a34-8783cc9fc7c9" path="/var/lib/kubelet/pods/b92605b2-50a6-41bc-8a34-8783cc9fc7c9/volumes" Oct 11 04:07:39 crc kubenswrapper[4967]: I1011 04:07:39.038200 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-56df8fb6b7-k8f6p" Oct 11 04:07:39 crc kubenswrapper[4967]: I1011 04:07:39.101718 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-58pw9"] Oct 11 04:07:39 crc kubenswrapper[4967]: I1011 04:07:39.101938 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-b8fbc5445-58pw9" podUID="a6cac387-4924-460d-8083-d9f4c296f2a3" containerName="dnsmasq-dns" containerID="cri-o://5f4ca90a58108c90e60b5b8d2945a4fbe89b9bfd3a1e06fb1885314e1f87da87" gracePeriod=10 Oct 11 04:07:39 crc kubenswrapper[4967]: I1011 04:07:39.287952 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-8389-account-create-pvbnb"] Oct 11 04:07:39 crc kubenswrapper[4967]: I1011 04:07:39.331229 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5772c259-389a-4faf-8307-f14eff408981","Type":"ContainerStarted","Data":"e2238c27782a7e923fc20d21eaa4482d40bf397b130fc23094305eb0b3309757"} Oct 11 04:07:39 crc kubenswrapper[4967]: I1011 04:07:39.333582 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 11 04:07:39 crc kubenswrapper[4967]: I1011 04:07:39.339892 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5dtlx" event={"ID":"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3","Type":"ContainerStarted","Data":"4b56fc23eeae7214925b7a9914c1b342181d9803ea7a698b33723c05b84b8292"} Oct 11 04:07:39 crc kubenswrapper[4967]: I1011 04:07:39.339939 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5dtlx" event={"ID":"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3","Type":"ContainerStarted","Data":"1b7755fa8167f6713e49b6c49a4efec8ecfab315ff19c233b953e11b3048e56b"} Oct 11 04:07:39 crc kubenswrapper[4967]: W1011 04:07:39.359266 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf3fcbf92_aff2_444c_abf1_b23c45ef96e9.slice/crio-54c57f77efaa0ea0680411e499b7a5aa95c6b8d5fea0a28e30ebfdfc07463f01 WatchSource:0}: Error finding container 54c57f77efaa0ea0680411e499b7a5aa95c6b8d5fea0a28e30ebfdfc07463f01: Status 404 returned error can't find the container with id 54c57f77efaa0ea0680411e499b7a5aa95c6b8d5fea0a28e30ebfdfc07463f01 Oct 11 04:07:39 crc kubenswrapper[4967]: I1011 04:07:39.799695 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-317e-account-create-l88bq" Oct 11 04:07:39 crc kubenswrapper[4967]: I1011 04:07:39.830950 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-bootstrap-5dtlx" podStartSLOduration=2.830933095 podStartE2EDuration="2.830933095s" podCreationTimestamp="2025-10-11 04:07:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:07:39.359195707 +0000 UTC m=+987.322404630" watchObservedRunningTime="2025-10-11 04:07:39.830933095 +0000 UTC m=+987.794142028" Oct 11 04:07:39 crc kubenswrapper[4967]: I1011 04:07:39.880181 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5b20-account-create-lb22f" Oct 11 04:07:39 crc kubenswrapper[4967]: I1011 04:07:39.896384 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-b879-account-create-6v2r9" Oct 11 04:07:39 crc kubenswrapper[4967]: I1011 04:07:39.934819 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-c5z8c\" (UniqueName: \"kubernetes.io/projected/2930d6b7-0aad-4123-b8bd-4dad4b8768fe-kube-api-access-c5z8c\") pod \"2930d6b7-0aad-4123-b8bd-4dad4b8768fe\" (UID: \"2930d6b7-0aad-4123-b8bd-4dad4b8768fe\") " Oct 11 04:07:39 crc kubenswrapper[4967]: I1011 04:07:39.942537 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2930d6b7-0aad-4123-b8bd-4dad4b8768fe-kube-api-access-c5z8c" (OuterVolumeSpecName: "kube-api-access-c5z8c") pod "2930d6b7-0aad-4123-b8bd-4dad4b8768fe" (UID: "2930d6b7-0aad-4123-b8bd-4dad4b8768fe"). InnerVolumeSpecName "kube-api-access-c5z8c". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:07:40 crc kubenswrapper[4967]: I1011 04:07:40.037128 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slch9\" (UniqueName: \"kubernetes.io/projected/92ccb4c4-e0b5-4592-bdfc-9909ee48b8b1-kube-api-access-slch9\") pod \"92ccb4c4-e0b5-4592-bdfc-9909ee48b8b1\" (UID: \"92ccb4c4-e0b5-4592-bdfc-9909ee48b8b1\") " Oct 11 04:07:40 crc kubenswrapper[4967]: I1011 04:07:40.037305 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lnv6d\" (UniqueName: \"kubernetes.io/projected/2743338e-1478-4402-b29f-d4b86722d241-kube-api-access-lnv6d\") pod \"2743338e-1478-4402-b29f-d4b86722d241\" (UID: \"2743338e-1478-4402-b29f-d4b86722d241\") " Oct 11 04:07:40 crc kubenswrapper[4967]: I1011 04:07:40.037879 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-c5z8c\" (UniqueName: \"kubernetes.io/projected/2930d6b7-0aad-4123-b8bd-4dad4b8768fe-kube-api-access-c5z8c\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:40 crc kubenswrapper[4967]: I1011 04:07:40.040056 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2743338e-1478-4402-b29f-d4b86722d241-kube-api-access-lnv6d" (OuterVolumeSpecName: "kube-api-access-lnv6d") pod "2743338e-1478-4402-b29f-d4b86722d241" (UID: "2743338e-1478-4402-b29f-d4b86722d241"). InnerVolumeSpecName "kube-api-access-lnv6d". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:07:40 crc kubenswrapper[4967]: I1011 04:07:40.043186 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92ccb4c4-e0b5-4592-bdfc-9909ee48b8b1-kube-api-access-slch9" (OuterVolumeSpecName: "kube-api-access-slch9") pod "92ccb4c4-e0b5-4592-bdfc-9909ee48b8b1" (UID: "92ccb4c4-e0b5-4592-bdfc-9909ee48b8b1"). InnerVolumeSpecName "kube-api-access-slch9". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:07:40 crc kubenswrapper[4967]: I1011 04:07:40.138912 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slch9\" (UniqueName: \"kubernetes.io/projected/92ccb4c4-e0b5-4592-bdfc-9909ee48b8b1-kube-api-access-slch9\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:40 crc kubenswrapper[4967]: I1011 04:07:40.138940 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lnv6d\" (UniqueName: \"kubernetes.io/projected/2743338e-1478-4402-b29f-d4b86722d241-kube-api-access-lnv6d\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:40 crc kubenswrapper[4967]: I1011 04:07:40.347359 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-b879-account-create-6v2r9" event={"ID":"92ccb4c4-e0b5-4592-bdfc-9909ee48b8b1","Type":"ContainerDied","Data":"c85053d0d5508ed96808221409a03c305dd73770f11fe1bbfe75ab4b7c89324d"} Oct 11 04:07:40 crc kubenswrapper[4967]: I1011 04:07:40.347737 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c85053d0d5508ed96808221409a03c305dd73770f11fe1bbfe75ab4b7c89324d" Oct 11 04:07:40 crc kubenswrapper[4967]: I1011 04:07:40.347390 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-b879-account-create-6v2r9" Oct 11 04:07:40 crc kubenswrapper[4967]: I1011 04:07:40.350541 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-317e-account-create-l88bq" event={"ID":"2930d6b7-0aad-4123-b8bd-4dad4b8768fe","Type":"ContainerDied","Data":"ed6ebe5c5273dec147dc0d9f8cdbdade6490dc57c902315056fea30bdc3edd32"} Oct 11 04:07:40 crc kubenswrapper[4967]: I1011 04:07:40.350581 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="ed6ebe5c5273dec147dc0d9f8cdbdade6490dc57c902315056fea30bdc3edd32" Oct 11 04:07:40 crc kubenswrapper[4967]: I1011 04:07:40.350564 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-317e-account-create-l88bq" Oct 11 04:07:40 crc kubenswrapper[4967]: I1011 04:07:40.352188 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-8389-account-create-pvbnb" event={"ID":"4691fb42-e6bb-4e1b-a4c2-1f65e1292d4d","Type":"ContainerStarted","Data":"a449ac509f5bb96d022ad2dfa16f5f18061d368fa9b0707814c600556b7dd895"} Oct 11 04:07:40 crc kubenswrapper[4967]: I1011 04:07:40.353860 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-5b20-account-create-lb22f" Oct 11 04:07:40 crc kubenswrapper[4967]: I1011 04:07:40.356186 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-5b20-account-create-lb22f" event={"ID":"2743338e-1478-4402-b29f-d4b86722d241","Type":"ContainerDied","Data":"9af6783eb9a6a545ddbd706d12233acccfe85b00f45bceda347ec168058969ae"} Oct 11 04:07:40 crc kubenswrapper[4967]: I1011 04:07:40.356218 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9af6783eb9a6a545ddbd706d12233acccfe85b00f45bceda347ec168058969ae" Oct 11 04:07:40 crc kubenswrapper[4967]: I1011 04:07:40.357741 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f3fcbf92-aff2-444c-abf1-b23c45ef96e9","Type":"ContainerStarted","Data":"54c57f77efaa0ea0680411e499b7a5aa95c6b8d5fea0a28e30ebfdfc07463f01"} Oct 11 04:07:40 crc kubenswrapper[4967]: I1011 04:07:40.359822 4967 generic.go:334] "Generic (PLEG): container finished" podID="a6cac387-4924-460d-8083-d9f4c296f2a3" containerID="5f4ca90a58108c90e60b5b8d2945a4fbe89b9bfd3a1e06fb1885314e1f87da87" exitCode=0 Oct 11 04:07:40 crc kubenswrapper[4967]: I1011 04:07:40.359908 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-58pw9" event={"ID":"a6cac387-4924-460d-8083-d9f4c296f2a3","Type":"ContainerDied","Data":"5f4ca90a58108c90e60b5b8d2945a4fbe89b9bfd3a1e06fb1885314e1f87da87"} Oct 11 04:07:42 crc kubenswrapper[4967]: I1011 04:07:42.014297 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/dnsmasq-dns-b8fbc5445-58pw9" podUID="a6cac387-4924-460d-8083-d9f4c296f2a3" containerName="dnsmasq-dns" probeResult="failure" output="dial tcp 10.217.0.116:5353: connect: connection refused" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.018268 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-db-sync-6bhhf"] Oct 11 04:07:43 crc kubenswrapper[4967]: E1011 04:07:43.018664 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92ccb4c4-e0b5-4592-bdfc-9909ee48b8b1" containerName="mariadb-account-create" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.018680 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="92ccb4c4-e0b5-4592-bdfc-9909ee48b8b1" containerName="mariadb-account-create" Oct 11 04:07:43 crc kubenswrapper[4967]: E1011 04:07:43.018699 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2743338e-1478-4402-b29f-d4b86722d241" containerName="mariadb-account-create" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.018706 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="2743338e-1478-4402-b29f-d4b86722d241" containerName="mariadb-account-create" Oct 11 04:07:43 crc kubenswrapper[4967]: E1011 04:07:43.018719 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2930d6b7-0aad-4123-b8bd-4dad4b8768fe" containerName="mariadb-account-create" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.018728 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="2930d6b7-0aad-4123-b8bd-4dad4b8768fe" containerName="mariadb-account-create" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.018905 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="2930d6b7-0aad-4123-b8bd-4dad4b8768fe" containerName="mariadb-account-create" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.018929 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="92ccb4c4-e0b5-4592-bdfc-9909ee48b8b1" containerName="mariadb-account-create" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.018955 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="2743338e-1478-4402-b29f-d4b86722d241" containerName="mariadb-account-create" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.019600 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-6bhhf" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.024966 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.025214 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.025340 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-24g9r" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.033441 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-6bhhf"] Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.187485 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-db-sync-wdc9t"] Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.188464 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-wdc9t" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.191186 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-mltgm" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.191770 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dq4ts\" (UniqueName: \"kubernetes.io/projected/1680354a-b333-4e21-9748-0062d78a479c-kube-api-access-dq4ts\") pod \"cinder-db-sync-6bhhf\" (UID: \"1680354a-b333-4e21-9748-0062d78a479c\") " pod="openstack/cinder-db-sync-6bhhf" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.191836 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.191859 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1680354a-b333-4e21-9748-0062d78a479c-combined-ca-bundle\") pod \"cinder-db-sync-6bhhf\" (UID: \"1680354a-b333-4e21-9748-0062d78a479c\") " pod="openstack/cinder-db-sync-6bhhf" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.191892 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1680354a-b333-4e21-9748-0062d78a479c-etc-machine-id\") pod \"cinder-db-sync-6bhhf\" (UID: \"1680354a-b333-4e21-9748-0062d78a479c\") " pod="openstack/cinder-db-sync-6bhhf" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.191983 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1680354a-b333-4e21-9748-0062d78a479c-config-data\") pod \"cinder-db-sync-6bhhf\" (UID: \"1680354a-b333-4e21-9748-0062d78a479c\") " pod="openstack/cinder-db-sync-6bhhf" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.192024 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1680354a-b333-4e21-9748-0062d78a479c-db-sync-config-data\") pod \"cinder-db-sync-6bhhf\" (UID: \"1680354a-b333-4e21-9748-0062d78a479c\") " pod="openstack/cinder-db-sync-6bhhf" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.192043 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1680354a-b333-4e21-9748-0062d78a479c-scripts\") pod \"cinder-db-sync-6bhhf\" (UID: \"1680354a-b333-4e21-9748-0062d78a479c\") " pod="openstack/cinder-db-sync-6bhhf" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.200117 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-wdc9t"] Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.292985 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-49m56\" (UniqueName: \"kubernetes.io/projected/8539e0a2-fecc-46f8-aa2f-a931a31cf17f-kube-api-access-49m56\") pod \"barbican-db-sync-wdc9t\" (UID: \"8539e0a2-fecc-46f8-aa2f-a931a31cf17f\") " pod="openstack/barbican-db-sync-wdc9t" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.293058 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8539e0a2-fecc-46f8-aa2f-a931a31cf17f-db-sync-config-data\") pod \"barbican-db-sync-wdc9t\" (UID: \"8539e0a2-fecc-46f8-aa2f-a931a31cf17f\") " pod="openstack/barbican-db-sync-wdc9t" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.293108 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1680354a-b333-4e21-9748-0062d78a479c-config-data\") pod \"cinder-db-sync-6bhhf\" (UID: \"1680354a-b333-4e21-9748-0062d78a479c\") " pod="openstack/cinder-db-sync-6bhhf" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.293147 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1680354a-b333-4e21-9748-0062d78a479c-db-sync-config-data\") pod \"cinder-db-sync-6bhhf\" (UID: \"1680354a-b333-4e21-9748-0062d78a479c\") " pod="openstack/cinder-db-sync-6bhhf" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.293165 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1680354a-b333-4e21-9748-0062d78a479c-scripts\") pod \"cinder-db-sync-6bhhf\" (UID: \"1680354a-b333-4e21-9748-0062d78a479c\") " pod="openstack/cinder-db-sync-6bhhf" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.293192 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8539e0a2-fecc-46f8-aa2f-a931a31cf17f-combined-ca-bundle\") pod \"barbican-db-sync-wdc9t\" (UID: \"8539e0a2-fecc-46f8-aa2f-a931a31cf17f\") " pod="openstack/barbican-db-sync-wdc9t" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.293220 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dq4ts\" (UniqueName: \"kubernetes.io/projected/1680354a-b333-4e21-9748-0062d78a479c-kube-api-access-dq4ts\") pod \"cinder-db-sync-6bhhf\" (UID: \"1680354a-b333-4e21-9748-0062d78a479c\") " pod="openstack/cinder-db-sync-6bhhf" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.293277 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1680354a-b333-4e21-9748-0062d78a479c-combined-ca-bundle\") pod \"cinder-db-sync-6bhhf\" (UID: \"1680354a-b333-4e21-9748-0062d78a479c\") " pod="openstack/cinder-db-sync-6bhhf" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.293297 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1680354a-b333-4e21-9748-0062d78a479c-etc-machine-id\") pod \"cinder-db-sync-6bhhf\" (UID: \"1680354a-b333-4e21-9748-0062d78a479c\") " pod="openstack/cinder-db-sync-6bhhf" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.293375 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1680354a-b333-4e21-9748-0062d78a479c-etc-machine-id\") pod \"cinder-db-sync-6bhhf\" (UID: \"1680354a-b333-4e21-9748-0062d78a479c\") " pod="openstack/cinder-db-sync-6bhhf" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.303528 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1680354a-b333-4e21-9748-0062d78a479c-scripts\") pod \"cinder-db-sync-6bhhf\" (UID: \"1680354a-b333-4e21-9748-0062d78a479c\") " pod="openstack/cinder-db-sync-6bhhf" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.304434 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1680354a-b333-4e21-9748-0062d78a479c-config-data\") pod \"cinder-db-sync-6bhhf\" (UID: \"1680354a-b333-4e21-9748-0062d78a479c\") " pod="openstack/cinder-db-sync-6bhhf" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.304995 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1680354a-b333-4e21-9748-0062d78a479c-combined-ca-bundle\") pod \"cinder-db-sync-6bhhf\" (UID: \"1680354a-b333-4e21-9748-0062d78a479c\") " pod="openstack/cinder-db-sync-6bhhf" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.316353 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-db-sync-nqrkv"] Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.318592 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-nqrkv" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.318765 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dq4ts\" (UniqueName: \"kubernetes.io/projected/1680354a-b333-4e21-9748-0062d78a479c-kube-api-access-dq4ts\") pod \"cinder-db-sync-6bhhf\" (UID: \"1680354a-b333-4e21-9748-0062d78a479c\") " pod="openstack/cinder-db-sync-6bhhf" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.323449 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.323463 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1680354a-b333-4e21-9748-0062d78a479c-db-sync-config-data\") pod \"cinder-db-sync-6bhhf\" (UID: \"1680354a-b333-4e21-9748-0062d78a479c\") " pod="openstack/cinder-db-sync-6bhhf" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.323606 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-kpsm7" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.323826 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.329032 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-nqrkv"] Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.340466 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-6bhhf" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.394270 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8539e0a2-fecc-46f8-aa2f-a931a31cf17f-combined-ca-bundle\") pod \"barbican-db-sync-wdc9t\" (UID: \"8539e0a2-fecc-46f8-aa2f-a931a31cf17f\") " pod="openstack/barbican-db-sync-wdc9t" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.394311 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/28309ac7-8bd7-425c-a2f8-1041f1749acb-config\") pod \"neutron-db-sync-nqrkv\" (UID: \"28309ac7-8bd7-425c-a2f8-1041f1749acb\") " pod="openstack/neutron-db-sync-nqrkv" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.394343 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28309ac7-8bd7-425c-a2f8-1041f1749acb-combined-ca-bundle\") pod \"neutron-db-sync-nqrkv\" (UID: \"28309ac7-8bd7-425c-a2f8-1041f1749acb\") " pod="openstack/neutron-db-sync-nqrkv" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.394391 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d5j2j\" (UniqueName: \"kubernetes.io/projected/28309ac7-8bd7-425c-a2f8-1041f1749acb-kube-api-access-d5j2j\") pod \"neutron-db-sync-nqrkv\" (UID: \"28309ac7-8bd7-425c-a2f8-1041f1749acb\") " pod="openstack/neutron-db-sync-nqrkv" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.394419 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-49m56\" (UniqueName: \"kubernetes.io/projected/8539e0a2-fecc-46f8-aa2f-a931a31cf17f-kube-api-access-49m56\") pod \"barbican-db-sync-wdc9t\" (UID: \"8539e0a2-fecc-46f8-aa2f-a931a31cf17f\") " pod="openstack/barbican-db-sync-wdc9t" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.394465 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8539e0a2-fecc-46f8-aa2f-a931a31cf17f-db-sync-config-data\") pod \"barbican-db-sync-wdc9t\" (UID: \"8539e0a2-fecc-46f8-aa2f-a931a31cf17f\") " pod="openstack/barbican-db-sync-wdc9t" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.398976 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8539e0a2-fecc-46f8-aa2f-a931a31cf17f-db-sync-config-data\") pod \"barbican-db-sync-wdc9t\" (UID: \"8539e0a2-fecc-46f8-aa2f-a931a31cf17f\") " pod="openstack/barbican-db-sync-wdc9t" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.405242 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8539e0a2-fecc-46f8-aa2f-a931a31cf17f-combined-ca-bundle\") pod \"barbican-db-sync-wdc9t\" (UID: \"8539e0a2-fecc-46f8-aa2f-a931a31cf17f\") " pod="openstack/barbican-db-sync-wdc9t" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.412560 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-49m56\" (UniqueName: \"kubernetes.io/projected/8539e0a2-fecc-46f8-aa2f-a931a31cf17f-kube-api-access-49m56\") pod \"barbican-db-sync-wdc9t\" (UID: \"8539e0a2-fecc-46f8-aa2f-a931a31cf17f\") " pod="openstack/barbican-db-sync-wdc9t" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.497184 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28309ac7-8bd7-425c-a2f8-1041f1749acb-combined-ca-bundle\") pod \"neutron-db-sync-nqrkv\" (UID: \"28309ac7-8bd7-425c-a2f8-1041f1749acb\") " pod="openstack/neutron-db-sync-nqrkv" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.497273 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d5j2j\" (UniqueName: \"kubernetes.io/projected/28309ac7-8bd7-425c-a2f8-1041f1749acb-kube-api-access-d5j2j\") pod \"neutron-db-sync-nqrkv\" (UID: \"28309ac7-8bd7-425c-a2f8-1041f1749acb\") " pod="openstack/neutron-db-sync-nqrkv" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.497403 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/28309ac7-8bd7-425c-a2f8-1041f1749acb-config\") pod \"neutron-db-sync-nqrkv\" (UID: \"28309ac7-8bd7-425c-a2f8-1041f1749acb\") " pod="openstack/neutron-db-sync-nqrkv" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.514495 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/28309ac7-8bd7-425c-a2f8-1041f1749acb-config\") pod \"neutron-db-sync-nqrkv\" (UID: \"28309ac7-8bd7-425c-a2f8-1041f1749acb\") " pod="openstack/neutron-db-sync-nqrkv" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.514815 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28309ac7-8bd7-425c-a2f8-1041f1749acb-combined-ca-bundle\") pod \"neutron-db-sync-nqrkv\" (UID: \"28309ac7-8bd7-425c-a2f8-1041f1749acb\") " pod="openstack/neutron-db-sync-nqrkv" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.514955 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-wdc9t" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.525291 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d5j2j\" (UniqueName: \"kubernetes.io/projected/28309ac7-8bd7-425c-a2f8-1041f1749acb-kube-api-access-d5j2j\") pod \"neutron-db-sync-nqrkv\" (UID: \"28309ac7-8bd7-425c-a2f8-1041f1749acb\") " pod="openstack/neutron-db-sync-nqrkv" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.795481 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-nqrkv" Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.856920 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-db-sync-6bhhf"] Oct 11 04:07:43 crc kubenswrapper[4967]: W1011 04:07:43.861054 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1680354a_b333_4e21_9748_0062d78a479c.slice/crio-7d6e1e33ad71585b4c706ec5ff964a445a52f317855abab60703336fc3333040 WatchSource:0}: Error finding container 7d6e1e33ad71585b4c706ec5ff964a445a52f317855abab60703336fc3333040: Status 404 returned error can't find the container with id 7d6e1e33ad71585b4c706ec5ff964a445a52f317855abab60703336fc3333040 Oct 11 04:07:43 crc kubenswrapper[4967]: I1011 04:07:43.971914 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-db-sync-wdc9t"] Oct 11 04:07:43 crc kubenswrapper[4967]: W1011 04:07:43.982934 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8539e0a2_fecc_46f8_aa2f_a931a31cf17f.slice/crio-3329e348c64cf3c0e690d218092c4f6adc2fbcfb3827fd51b3654641237bcf49 WatchSource:0}: Error finding container 3329e348c64cf3c0e690d218092c4f6adc2fbcfb3827fd51b3654641237bcf49: Status 404 returned error can't find the container with id 3329e348c64cf3c0e690d218092c4f6adc2fbcfb3827fd51b3654641237bcf49 Oct 11 04:07:44 crc kubenswrapper[4967]: I1011 04:07:44.258131 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-db-sync-nqrkv"] Oct 11 04:07:44 crc kubenswrapper[4967]: W1011 04:07:44.266787 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod28309ac7_8bd7_425c_a2f8_1041f1749acb.slice/crio-b62926e006d4a5758e50d894276bb514a52435eebbc618847cae44df89e55bd9 WatchSource:0}: Error finding container b62926e006d4a5758e50d894276bb514a52435eebbc618847cae44df89e55bd9: Status 404 returned error can't find the container with id b62926e006d4a5758e50d894276bb514a52435eebbc618847cae44df89e55bd9 Oct 11 04:07:44 crc kubenswrapper[4967]: I1011 04:07:44.397483 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-nqrkv" event={"ID":"28309ac7-8bd7-425c-a2f8-1041f1749acb","Type":"ContainerStarted","Data":"b62926e006d4a5758e50d894276bb514a52435eebbc618847cae44df89e55bd9"} Oct 11 04:07:44 crc kubenswrapper[4967]: I1011 04:07:44.405406 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-8389-account-create-pvbnb" event={"ID":"4691fb42-e6bb-4e1b-a4c2-1f65e1292d4d","Type":"ContainerStarted","Data":"208f95e3e45952d983d55b0410f07c0a8a43bff9ada9afb6bcc1da3bf5f1f71e"} Oct 11 04:07:44 crc kubenswrapper[4967]: I1011 04:07:44.429977 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f3fcbf92-aff2-444c-abf1-b23c45ef96e9","Type":"ContainerStarted","Data":"0e68c39cb6ce4b2246948b7da0836b7edf0d64a7d6f59868bf31e93d2857f817"} Oct 11 04:07:44 crc kubenswrapper[4967]: I1011 04:07:44.444963 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-8389-account-create-pvbnb" podStartSLOduration=6.444945125 podStartE2EDuration="6.444945125s" podCreationTimestamp="2025-10-11 04:07:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:07:44.421822361 +0000 UTC m=+992.385031294" watchObservedRunningTime="2025-10-11 04:07:44.444945125 +0000 UTC m=+992.408154058" Oct 11 04:07:44 crc kubenswrapper[4967]: I1011 04:07:44.450524 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5772c259-389a-4faf-8307-f14eff408981","Type":"ContainerStarted","Data":"6b48779a5bbe12fa36091acda9cf3f4bca4678cb161b66dd0ed62b30620a5c7c"} Oct 11 04:07:44 crc kubenswrapper[4967]: I1011 04:07:44.457187 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-wdc9t" event={"ID":"8539e0a2-fecc-46f8-aa2f-a931a31cf17f","Type":"ContainerStarted","Data":"3329e348c64cf3c0e690d218092c4f6adc2fbcfb3827fd51b3654641237bcf49"} Oct 11 04:07:44 crc kubenswrapper[4967]: I1011 04:07:44.469220 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-6bhhf" event={"ID":"1680354a-b333-4e21-9748-0062d78a479c","Type":"ContainerStarted","Data":"7d6e1e33ad71585b4c706ec5ff964a445a52f317855abab60703336fc3333040"} Oct 11 04:07:44 crc kubenswrapper[4967]: I1011 04:07:44.738483 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-58pw9" Oct 11 04:07:44 crc kubenswrapper[4967]: I1011 04:07:44.830999 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a6cac387-4924-460d-8083-d9f4c296f2a3-ovsdbserver-sb\") pod \"a6cac387-4924-460d-8083-d9f4c296f2a3\" (UID: \"a6cac387-4924-460d-8083-d9f4c296f2a3\") " Oct 11 04:07:44 crc kubenswrapper[4967]: I1011 04:07:44.831436 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a6cac387-4924-460d-8083-d9f4c296f2a3-ovsdbserver-nb\") pod \"a6cac387-4924-460d-8083-d9f4c296f2a3\" (UID: \"a6cac387-4924-460d-8083-d9f4c296f2a3\") " Oct 11 04:07:44 crc kubenswrapper[4967]: I1011 04:07:44.831615 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6cac387-4924-460d-8083-d9f4c296f2a3-config\") pod \"a6cac387-4924-460d-8083-d9f4c296f2a3\" (UID: \"a6cac387-4924-460d-8083-d9f4c296f2a3\") " Oct 11 04:07:44 crc kubenswrapper[4967]: I1011 04:07:44.831659 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6cac387-4924-460d-8083-d9f4c296f2a3-dns-svc\") pod \"a6cac387-4924-460d-8083-d9f4c296f2a3\" (UID: \"a6cac387-4924-460d-8083-d9f4c296f2a3\") " Oct 11 04:07:44 crc kubenswrapper[4967]: I1011 04:07:44.831677 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d2v9n\" (UniqueName: \"kubernetes.io/projected/a6cac387-4924-460d-8083-d9f4c296f2a3-kube-api-access-d2v9n\") pod \"a6cac387-4924-460d-8083-d9f4c296f2a3\" (UID: \"a6cac387-4924-460d-8083-d9f4c296f2a3\") " Oct 11 04:07:44 crc kubenswrapper[4967]: I1011 04:07:44.836390 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a6cac387-4924-460d-8083-d9f4c296f2a3-kube-api-access-d2v9n" (OuterVolumeSpecName: "kube-api-access-d2v9n") pod "a6cac387-4924-460d-8083-d9f4c296f2a3" (UID: "a6cac387-4924-460d-8083-d9f4c296f2a3"). InnerVolumeSpecName "kube-api-access-d2v9n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:07:44 crc kubenswrapper[4967]: I1011 04:07:44.894572 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6cac387-4924-460d-8083-d9f4c296f2a3-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "a6cac387-4924-460d-8083-d9f4c296f2a3" (UID: "a6cac387-4924-460d-8083-d9f4c296f2a3"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:07:44 crc kubenswrapper[4967]: I1011 04:07:44.896508 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6cac387-4924-460d-8083-d9f4c296f2a3-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "a6cac387-4924-460d-8083-d9f4c296f2a3" (UID: "a6cac387-4924-460d-8083-d9f4c296f2a3"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:07:44 crc kubenswrapper[4967]: I1011 04:07:44.896664 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6cac387-4924-460d-8083-d9f4c296f2a3-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "a6cac387-4924-460d-8083-d9f4c296f2a3" (UID: "a6cac387-4924-460d-8083-d9f4c296f2a3"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:07:44 crc kubenswrapper[4967]: I1011 04:07:44.897049 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a6cac387-4924-460d-8083-d9f4c296f2a3-config" (OuterVolumeSpecName: "config") pod "a6cac387-4924-460d-8083-d9f4c296f2a3" (UID: "a6cac387-4924-460d-8083-d9f4c296f2a3"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:07:44 crc kubenswrapper[4967]: I1011 04:07:44.933488 4967 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/a6cac387-4924-460d-8083-d9f4c296f2a3-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:44 crc kubenswrapper[4967]: I1011 04:07:44.933521 4967 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/a6cac387-4924-460d-8083-d9f4c296f2a3-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:44 crc kubenswrapper[4967]: I1011 04:07:44.933533 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6cac387-4924-460d-8083-d9f4c296f2a3-config\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:44 crc kubenswrapper[4967]: I1011 04:07:44.933544 4967 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/a6cac387-4924-460d-8083-d9f4c296f2a3-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:44 crc kubenswrapper[4967]: I1011 04:07:44.933554 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d2v9n\" (UniqueName: \"kubernetes.io/projected/a6cac387-4924-460d-8083-d9f4c296f2a3-kube-api-access-d2v9n\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:45 crc kubenswrapper[4967]: I1011 04:07:45.480848 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-b8fbc5445-58pw9" Oct 11 04:07:45 crc kubenswrapper[4967]: I1011 04:07:45.480858 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-b8fbc5445-58pw9" event={"ID":"a6cac387-4924-460d-8083-d9f4c296f2a3","Type":"ContainerDied","Data":"d4944dab1f7db267d8aeebad3afb684f9385329957f3e65838fcc251386f1c72"} Oct 11 04:07:45 crc kubenswrapper[4967]: I1011 04:07:45.482101 4967 scope.go:117] "RemoveContainer" containerID="5f4ca90a58108c90e60b5b8d2945a4fbe89b9bfd3a1e06fb1885314e1f87da87" Oct 11 04:07:45 crc kubenswrapper[4967]: I1011 04:07:45.484134 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5772c259-389a-4faf-8307-f14eff408981","Type":"ContainerStarted","Data":"7aa3f6cc45ece2e90b3cbb12c8c263e2c384fb23f3130d0e9106fedcf0f3b527"} Oct 11 04:07:45 crc kubenswrapper[4967]: I1011 04:07:45.484331 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5772c259-389a-4faf-8307-f14eff408981" containerName="glance-log" containerID="cri-o://6b48779a5bbe12fa36091acda9cf3f4bca4678cb161b66dd0ed62b30620a5c7c" gracePeriod=30 Oct 11 04:07:45 crc kubenswrapper[4967]: I1011 04:07:45.484367 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="5772c259-389a-4faf-8307-f14eff408981" containerName="glance-httpd" containerID="cri-o://7aa3f6cc45ece2e90b3cbb12c8c263e2c384fb23f3130d0e9106fedcf0f3b527" gracePeriod=30 Oct 11 04:07:45 crc kubenswrapper[4967]: I1011 04:07:45.486578 4967 generic.go:334] "Generic (PLEG): container finished" podID="05059c3c-e29a-4bfa-a0d4-0d56b3d85af9" containerID="abdc71002731f1f79b9c3e24a5bce139cf8e8ef69665b9d7a89453b7eda936ce" exitCode=0 Oct 11 04:07:45 crc kubenswrapper[4967]: I1011 04:07:45.486764 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zf6xs" event={"ID":"05059c3c-e29a-4bfa-a0d4-0d56b3d85af9","Type":"ContainerDied","Data":"abdc71002731f1f79b9c3e24a5bce139cf8e8ef69665b9d7a89453b7eda936ce"} Oct 11 04:07:45 crc kubenswrapper[4967]: I1011 04:07:45.489225 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-nqrkv" event={"ID":"28309ac7-8bd7-425c-a2f8-1041f1749acb","Type":"ContainerStarted","Data":"4f4319ea2749661e591abb39022c4bacf81f997235ba5a04b06afb53e61abb94"} Oct 11 04:07:45 crc kubenswrapper[4967]: I1011 04:07:45.495693 4967 generic.go:334] "Generic (PLEG): container finished" podID="dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3" containerID="4b56fc23eeae7214925b7a9914c1b342181d9803ea7a698b33723c05b84b8292" exitCode=0 Oct 11 04:07:45 crc kubenswrapper[4967]: I1011 04:07:45.495856 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5dtlx" event={"ID":"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3","Type":"ContainerDied","Data":"4b56fc23eeae7214925b7a9914c1b342181d9803ea7a698b33723c05b84b8292"} Oct 11 04:07:45 crc kubenswrapper[4967]: I1011 04:07:45.497979 4967 generic.go:334] "Generic (PLEG): container finished" podID="4691fb42-e6bb-4e1b-a4c2-1f65e1292d4d" containerID="208f95e3e45952d983d55b0410f07c0a8a43bff9ada9afb6bcc1da3bf5f1f71e" exitCode=0 Oct 11 04:07:45 crc kubenswrapper[4967]: I1011 04:07:45.498123 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-8389-account-create-pvbnb" event={"ID":"4691fb42-e6bb-4e1b-a4c2-1f65e1292d4d","Type":"ContainerDied","Data":"208f95e3e45952d983d55b0410f07c0a8a43bff9ada9afb6bcc1da3bf5f1f71e"} Oct 11 04:07:45 crc kubenswrapper[4967]: I1011 04:07:45.504781 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb56ccd5-8cd3-4da0-b11f-8600af8abf86","Type":"ContainerStarted","Data":"ec8b91f055b4e4c15ad8ebb839b6d61ae69225f0fd74f6083d4bc346cf2ab43b"} Oct 11 04:07:45 crc kubenswrapper[4967]: I1011 04:07:45.515964 4967 scope.go:117] "RemoveContainer" containerID="7eb4833878553650843b57cb3dbfd20b888d8a902999cbe64d1f71b82df556b5" Oct 11 04:07:45 crc kubenswrapper[4967]: I1011 04:07:45.544040 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=8.544014109999999 podStartE2EDuration="8.54401411s" podCreationTimestamp="2025-10-11 04:07:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:07:45.536502796 +0000 UTC m=+993.499711739" watchObservedRunningTime="2025-10-11 04:07:45.54401411 +0000 UTC m=+993.507223073" Oct 11 04:07:45 crc kubenswrapper[4967]: I1011 04:07:45.559569 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-db-sync-nqrkv" podStartSLOduration=2.559538148 podStartE2EDuration="2.559538148s" podCreationTimestamp="2025-10-11 04:07:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:07:45.558262128 +0000 UTC m=+993.521471081" watchObservedRunningTime="2025-10-11 04:07:45.559538148 +0000 UTC m=+993.522747121" Oct 11 04:07:45 crc kubenswrapper[4967]: I1011 04:07:45.589912 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-58pw9"] Oct 11 04:07:45 crc kubenswrapper[4967]: I1011 04:07:45.604355 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-b8fbc5445-58pw9"] Oct 11 04:07:46 crc kubenswrapper[4967]: I1011 04:07:46.517457 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f3fcbf92-aff2-444c-abf1-b23c45ef96e9","Type":"ContainerStarted","Data":"22aca31c65aece17da7d51c55636940bceadde7fd9c8f3b5c5066057250f4000"} Oct 11 04:07:46 crc kubenswrapper[4967]: I1011 04:07:46.519657 4967 generic.go:334] "Generic (PLEG): container finished" podID="5772c259-389a-4faf-8307-f14eff408981" containerID="7aa3f6cc45ece2e90b3cbb12c8c263e2c384fb23f3130d0e9106fedcf0f3b527" exitCode=0 Oct 11 04:07:46 crc kubenswrapper[4967]: I1011 04:07:46.519687 4967 generic.go:334] "Generic (PLEG): container finished" podID="5772c259-389a-4faf-8307-f14eff408981" containerID="6b48779a5bbe12fa36091acda9cf3f4bca4678cb161b66dd0ed62b30620a5c7c" exitCode=143 Oct 11 04:07:46 crc kubenswrapper[4967]: I1011 04:07:46.519722 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5772c259-389a-4faf-8307-f14eff408981","Type":"ContainerDied","Data":"7aa3f6cc45ece2e90b3cbb12c8c263e2c384fb23f3130d0e9106fedcf0f3b527"} Oct 11 04:07:46 crc kubenswrapper[4967]: I1011 04:07:46.519745 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5772c259-389a-4faf-8307-f14eff408981","Type":"ContainerDied","Data":"6b48779a5bbe12fa36091acda9cf3f4bca4678cb161b66dd0ed62b30620a5c7c"} Oct 11 04:07:46 crc kubenswrapper[4967]: I1011 04:07:46.550832 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=8.550816297 podStartE2EDuration="8.550816297s" podCreationTimestamp="2025-10-11 04:07:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:07:46.540251933 +0000 UTC m=+994.503460866" watchObservedRunningTime="2025-10-11 04:07:46.550816297 +0000 UTC m=+994.514025230" Oct 11 04:07:46 crc kubenswrapper[4967]: I1011 04:07:46.658205 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 11 04:07:46 crc kubenswrapper[4967]: I1011 04:07:46.793460 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5772c259-389a-4faf-8307-f14eff408981-logs\") pod \"5772c259-389a-4faf-8307-f14eff408981\" (UID: \"5772c259-389a-4faf-8307-f14eff408981\") " Oct 11 04:07:46 crc kubenswrapper[4967]: I1011 04:07:46.793936 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6cxvs\" (UniqueName: \"kubernetes.io/projected/5772c259-389a-4faf-8307-f14eff408981-kube-api-access-6cxvs\") pod \"5772c259-389a-4faf-8307-f14eff408981\" (UID: \"5772c259-389a-4faf-8307-f14eff408981\") " Oct 11 04:07:46 crc kubenswrapper[4967]: I1011 04:07:46.793972 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"5772c259-389a-4faf-8307-f14eff408981\" (UID: \"5772c259-389a-4faf-8307-f14eff408981\") " Oct 11 04:07:46 crc kubenswrapper[4967]: I1011 04:07:46.794034 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5772c259-389a-4faf-8307-f14eff408981-config-data\") pod \"5772c259-389a-4faf-8307-f14eff408981\" (UID: \"5772c259-389a-4faf-8307-f14eff408981\") " Oct 11 04:07:46 crc kubenswrapper[4967]: I1011 04:07:46.794201 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5772c259-389a-4faf-8307-f14eff408981-scripts\") pod \"5772c259-389a-4faf-8307-f14eff408981\" (UID: \"5772c259-389a-4faf-8307-f14eff408981\") " Oct 11 04:07:46 crc kubenswrapper[4967]: I1011 04:07:46.794255 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5772c259-389a-4faf-8307-f14eff408981-combined-ca-bundle\") pod \"5772c259-389a-4faf-8307-f14eff408981\" (UID: \"5772c259-389a-4faf-8307-f14eff408981\") " Oct 11 04:07:46 crc kubenswrapper[4967]: I1011 04:07:46.794351 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5772c259-389a-4faf-8307-f14eff408981-httpd-run\") pod \"5772c259-389a-4faf-8307-f14eff408981\" (UID: \"5772c259-389a-4faf-8307-f14eff408981\") " Oct 11 04:07:46 crc kubenswrapper[4967]: I1011 04:07:46.794917 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5772c259-389a-4faf-8307-f14eff408981-logs" (OuterVolumeSpecName: "logs") pod "5772c259-389a-4faf-8307-f14eff408981" (UID: "5772c259-389a-4faf-8307-f14eff408981"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:07:46 crc kubenswrapper[4967]: I1011 04:07:46.796766 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5772c259-389a-4faf-8307-f14eff408981-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "5772c259-389a-4faf-8307-f14eff408981" (UID: "5772c259-389a-4faf-8307-f14eff408981"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:07:46 crc kubenswrapper[4967]: I1011 04:07:46.800269 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "5772c259-389a-4faf-8307-f14eff408981" (UID: "5772c259-389a-4faf-8307-f14eff408981"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 11 04:07:46 crc kubenswrapper[4967]: I1011 04:07:46.806191 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5772c259-389a-4faf-8307-f14eff408981-kube-api-access-6cxvs" (OuterVolumeSpecName: "kube-api-access-6cxvs") pod "5772c259-389a-4faf-8307-f14eff408981" (UID: "5772c259-389a-4faf-8307-f14eff408981"). InnerVolumeSpecName "kube-api-access-6cxvs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:07:46 crc kubenswrapper[4967]: I1011 04:07:46.806270 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5772c259-389a-4faf-8307-f14eff408981-scripts" (OuterVolumeSpecName: "scripts") pod "5772c259-389a-4faf-8307-f14eff408981" (UID: "5772c259-389a-4faf-8307-f14eff408981"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:07:46 crc kubenswrapper[4967]: I1011 04:07:46.846185 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a6cac387-4924-460d-8083-d9f4c296f2a3" path="/var/lib/kubelet/pods/a6cac387-4924-460d-8083-d9f4c296f2a3/volumes" Oct 11 04:07:46 crc kubenswrapper[4967]: I1011 04:07:46.863219 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5772c259-389a-4faf-8307-f14eff408981-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5772c259-389a-4faf-8307-f14eff408981" (UID: "5772c259-389a-4faf-8307-f14eff408981"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:07:46 crc kubenswrapper[4967]: I1011 04:07:46.863329 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5772c259-389a-4faf-8307-f14eff408981-config-data" (OuterVolumeSpecName: "config-data") pod "5772c259-389a-4faf-8307-f14eff408981" (UID: "5772c259-389a-4faf-8307-f14eff408981"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:07:46 crc kubenswrapper[4967]: I1011 04:07:46.899924 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5772c259-389a-4faf-8307-f14eff408981-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:46 crc kubenswrapper[4967]: I1011 04:07:46.899996 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5772c259-389a-4faf-8307-f14eff408981-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:46 crc kubenswrapper[4967]: I1011 04:07:46.900009 4967 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/5772c259-389a-4faf-8307-f14eff408981-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:46 crc kubenswrapper[4967]: I1011 04:07:46.900018 4967 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5772c259-389a-4faf-8307-f14eff408981-logs\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:46 crc kubenswrapper[4967]: I1011 04:07:46.900027 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6cxvs\" (UniqueName: \"kubernetes.io/projected/5772c259-389a-4faf-8307-f14eff408981-kube-api-access-6cxvs\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:46 crc kubenswrapper[4967]: I1011 04:07:46.900050 4967 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 11 04:07:46 crc kubenswrapper[4967]: I1011 04:07:46.900058 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5772c259-389a-4faf-8307-f14eff408981-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:46 crc kubenswrapper[4967]: I1011 04:07:46.921885 4967 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 11 04:07:46 crc kubenswrapper[4967]: I1011 04:07:46.937508 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5dtlx" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.003020 4967 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.051142 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-8389-account-create-pvbnb" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.063809 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zf6xs" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.104442 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-scripts\") pod \"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3\" (UID: \"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3\") " Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.105048 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-config-data\") pod \"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3\" (UID: \"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3\") " Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.105116 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bp58w\" (UniqueName: \"kubernetes.io/projected/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-kube-api-access-bp58w\") pod \"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3\" (UID: \"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3\") " Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.105233 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-fernet-keys\") pod \"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3\" (UID: \"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3\") " Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.105260 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-credential-keys\") pod \"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3\" (UID: \"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3\") " Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.105304 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-combined-ca-bundle\") pod \"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3\" (UID: \"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3\") " Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.108324 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-scripts" (OuterVolumeSpecName: "scripts") pod "dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3" (UID: "dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.108494 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-kube-api-access-bp58w" (OuterVolumeSpecName: "kube-api-access-bp58w") pod "dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3" (UID: "dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3"). InnerVolumeSpecName "kube-api-access-bp58w". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.114224 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-credential-keys" (OuterVolumeSpecName: "credential-keys") pod "dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3" (UID: "dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3"). InnerVolumeSpecName "credential-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.114315 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-fernet-keys" (OuterVolumeSpecName: "fernet-keys") pod "dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3" (UID: "dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3"). InnerVolumeSpecName "fernet-keys". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.148405 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-config-data" (OuterVolumeSpecName: "config-data") pod "dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3" (UID: "dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.148721 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3" (UID: "dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.206903 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-272zg\" (UniqueName: \"kubernetes.io/projected/4691fb42-e6bb-4e1b-a4c2-1f65e1292d4d-kube-api-access-272zg\") pod \"4691fb42-e6bb-4e1b-a4c2-1f65e1292d4d\" (UID: \"4691fb42-e6bb-4e1b-a4c2-1f65e1292d4d\") " Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.206938 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w5r2v\" (UniqueName: \"kubernetes.io/projected/05059c3c-e29a-4bfa-a0d4-0d56b3d85af9-kube-api-access-w5r2v\") pod \"05059c3c-e29a-4bfa-a0d4-0d56b3d85af9\" (UID: \"05059c3c-e29a-4bfa-a0d4-0d56b3d85af9\") " Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.207002 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05059c3c-e29a-4bfa-a0d4-0d56b3d85af9-logs\") pod \"05059c3c-e29a-4bfa-a0d4-0d56b3d85af9\" (UID: \"05059c3c-e29a-4bfa-a0d4-0d56b3d85af9\") " Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.207023 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05059c3c-e29a-4bfa-a0d4-0d56b3d85af9-combined-ca-bundle\") pod \"05059c3c-e29a-4bfa-a0d4-0d56b3d85af9\" (UID: \"05059c3c-e29a-4bfa-a0d4-0d56b3d85af9\") " Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.207062 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05059c3c-e29a-4bfa-a0d4-0d56b3d85af9-scripts\") pod \"05059c3c-e29a-4bfa-a0d4-0d56b3d85af9\" (UID: \"05059c3c-e29a-4bfa-a0d4-0d56b3d85af9\") " Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.207096 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05059c3c-e29a-4bfa-a0d4-0d56b3d85af9-config-data\") pod \"05059c3c-e29a-4bfa-a0d4-0d56b3d85af9\" (UID: \"05059c3c-e29a-4bfa-a0d4-0d56b3d85af9\") " Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.207401 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/05059c3c-e29a-4bfa-a0d4-0d56b3d85af9-logs" (OuterVolumeSpecName: "logs") pod "05059c3c-e29a-4bfa-a0d4-0d56b3d85af9" (UID: "05059c3c-e29a-4bfa-a0d4-0d56b3d85af9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.207434 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.207446 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.207456 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bp58w\" (UniqueName: \"kubernetes.io/projected/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-kube-api-access-bp58w\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.207486 4967 reconciler_common.go:293] "Volume detached for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-fernet-keys\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.207495 4967 reconciler_common.go:293] "Volume detached for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-credential-keys\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.207504 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.211145 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05059c3c-e29a-4bfa-a0d4-0d56b3d85af9-scripts" (OuterVolumeSpecName: "scripts") pod "05059c3c-e29a-4bfa-a0d4-0d56b3d85af9" (UID: "05059c3c-e29a-4bfa-a0d4-0d56b3d85af9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.211750 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4691fb42-e6bb-4e1b-a4c2-1f65e1292d4d-kube-api-access-272zg" (OuterVolumeSpecName: "kube-api-access-272zg") pod "4691fb42-e6bb-4e1b-a4c2-1f65e1292d4d" (UID: "4691fb42-e6bb-4e1b-a4c2-1f65e1292d4d"). InnerVolumeSpecName "kube-api-access-272zg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.213209 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/05059c3c-e29a-4bfa-a0d4-0d56b3d85af9-kube-api-access-w5r2v" (OuterVolumeSpecName: "kube-api-access-w5r2v") pod "05059c3c-e29a-4bfa-a0d4-0d56b3d85af9" (UID: "05059c3c-e29a-4bfa-a0d4-0d56b3d85af9"). InnerVolumeSpecName "kube-api-access-w5r2v". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.233232 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05059c3c-e29a-4bfa-a0d4-0d56b3d85af9-config-data" (OuterVolumeSpecName: "config-data") pod "05059c3c-e29a-4bfa-a0d4-0d56b3d85af9" (UID: "05059c3c-e29a-4bfa-a0d4-0d56b3d85af9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.234210 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/05059c3c-e29a-4bfa-a0d4-0d56b3d85af9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "05059c3c-e29a-4bfa-a0d4-0d56b3d85af9" (UID: "05059c3c-e29a-4bfa-a0d4-0d56b3d85af9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.309196 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-272zg\" (UniqueName: \"kubernetes.io/projected/4691fb42-e6bb-4e1b-a4c2-1f65e1292d4d-kube-api-access-272zg\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.309484 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w5r2v\" (UniqueName: \"kubernetes.io/projected/05059c3c-e29a-4bfa-a0d4-0d56b3d85af9-kube-api-access-w5r2v\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.309497 4967 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/05059c3c-e29a-4bfa-a0d4-0d56b3d85af9-logs\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.309511 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/05059c3c-e29a-4bfa-a0d4-0d56b3d85af9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.309526 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/05059c3c-e29a-4bfa-a0d4-0d56b3d85af9-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.309537 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/05059c3c-e29a-4bfa-a0d4-0d56b3d85af9-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.569685 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-bootstrap-5dtlx" event={"ID":"dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3","Type":"ContainerDied","Data":"1b7755fa8167f6713e49b6c49a4efec8ecfab315ff19c233b953e11b3048e56b"} Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.569737 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1b7755fa8167f6713e49b6c49a4efec8ecfab315ff19c233b953e11b3048e56b" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.569824 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-bootstrap-5dtlx" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.602567 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-8389-account-create-pvbnb" event={"ID":"4691fb42-e6bb-4e1b-a4c2-1f65e1292d4d","Type":"ContainerDied","Data":"a449ac509f5bb96d022ad2dfa16f5f18061d368fa9b0707814c600556b7dd895"} Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.602604 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a449ac509f5bb96d022ad2dfa16f5f18061d368fa9b0707814c600556b7dd895" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.602580 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-8389-account-create-pvbnb" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.612688 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/placement-5cdcf9b996-625cn"] Oct 11 04:07:47 crc kubenswrapper[4967]: E1011 04:07:47.613267 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5772c259-389a-4faf-8307-f14eff408981" containerName="glance-httpd" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.613290 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="5772c259-389a-4faf-8307-f14eff408981" containerName="glance-httpd" Oct 11 04:07:47 crc kubenswrapper[4967]: E1011 04:07:47.613324 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6cac387-4924-460d-8083-d9f4c296f2a3" containerName="init" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.613343 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6cac387-4924-460d-8083-d9f4c296f2a3" containerName="init" Oct 11 04:07:47 crc kubenswrapper[4967]: E1011 04:07:47.613405 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a6cac387-4924-460d-8083-d9f4c296f2a3" containerName="dnsmasq-dns" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.613418 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="a6cac387-4924-460d-8083-d9f4c296f2a3" containerName="dnsmasq-dns" Oct 11 04:07:47 crc kubenswrapper[4967]: E1011 04:07:47.613439 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4691fb42-e6bb-4e1b-a4c2-1f65e1292d4d" containerName="mariadb-account-create" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.613452 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="4691fb42-e6bb-4e1b-a4c2-1f65e1292d4d" containerName="mariadb-account-create" Oct 11 04:07:47 crc kubenswrapper[4967]: E1011 04:07:47.613482 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3" containerName="keystone-bootstrap" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.613494 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3" containerName="keystone-bootstrap" Oct 11 04:07:47 crc kubenswrapper[4967]: E1011 04:07:47.613519 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5772c259-389a-4faf-8307-f14eff408981" containerName="glance-log" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.613533 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="5772c259-389a-4faf-8307-f14eff408981" containerName="glance-log" Oct 11 04:07:47 crc kubenswrapper[4967]: E1011 04:07:47.613550 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="05059c3c-e29a-4bfa-a0d4-0d56b3d85af9" containerName="placement-db-sync" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.615394 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.617280 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="05059c3c-e29a-4bfa-a0d4-0d56b3d85af9" containerName="placement-db-sync" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.617574 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="5772c259-389a-4faf-8307-f14eff408981" containerName="glance-httpd" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.617589 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3" containerName="keystone-bootstrap" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.618116 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="4691fb42-e6bb-4e1b-a4c2-1f65e1292d4d" containerName="mariadb-account-create" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.618135 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="a6cac387-4924-460d-8083-d9f4c296f2a3" containerName="dnsmasq-dns" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.618156 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="5772c259-389a-4faf-8307-f14eff408981" containerName="glance-log" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.618165 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="05059c3c-e29a-4bfa-a0d4-0d56b3d85af9" containerName="placement-db-sync" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.620452 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"5772c259-389a-4faf-8307-f14eff408981","Type":"ContainerDied","Data":"e2238c27782a7e923fc20d21eaa4482d40bf397b130fc23094305eb0b3309757"} Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.620496 4967 scope.go:117] "RemoveContainer" containerID="7aa3f6cc45ece2e90b3cbb12c8c263e2c384fb23f3130d0e9106fedcf0f3b527" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.623608 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5cdcf9b996-625cn" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.627210 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5cdcf9b996-625cn"] Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.627365 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-public-svc" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.627361 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-placement-internal-svc" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.635600 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/placement-db-sync-zf6xs" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.635654 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-db-sync-zf6xs" event={"ID":"05059c3c-e29a-4bfa-a0d4-0d56b3d85af9","Type":"ContainerDied","Data":"2174c3ec42975523ddd78c08814d157c30d6a60c2699ff360bf600e540168733"} Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.635674 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2174c3ec42975523ddd78c08814d157c30d6a60c2699ff360bf600e540168733" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.641566 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/878b0a76-9c37-4210-870b-5fed33721831-public-tls-certs\") pod \"placement-5cdcf9b996-625cn\" (UID: \"878b0a76-9c37-4210-870b-5fed33721831\") " pod="openstack/placement-5cdcf9b996-625cn" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.641618 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/878b0a76-9c37-4210-870b-5fed33721831-internal-tls-certs\") pod \"placement-5cdcf9b996-625cn\" (UID: \"878b0a76-9c37-4210-870b-5fed33721831\") " pod="openstack/placement-5cdcf9b996-625cn" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.641860 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6fhn\" (UniqueName: \"kubernetes.io/projected/878b0a76-9c37-4210-870b-5fed33721831-kube-api-access-s6fhn\") pod \"placement-5cdcf9b996-625cn\" (UID: \"878b0a76-9c37-4210-870b-5fed33721831\") " pod="openstack/placement-5cdcf9b996-625cn" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.641914 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/878b0a76-9c37-4210-870b-5fed33721831-config-data\") pod \"placement-5cdcf9b996-625cn\" (UID: \"878b0a76-9c37-4210-870b-5fed33721831\") " pod="openstack/placement-5cdcf9b996-625cn" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.641994 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/878b0a76-9c37-4210-870b-5fed33721831-logs\") pod \"placement-5cdcf9b996-625cn\" (UID: \"878b0a76-9c37-4210-870b-5fed33721831\") " pod="openstack/placement-5cdcf9b996-625cn" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.642031 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/878b0a76-9c37-4210-870b-5fed33721831-scripts\") pod \"placement-5cdcf9b996-625cn\" (UID: \"878b0a76-9c37-4210-870b-5fed33721831\") " pod="openstack/placement-5cdcf9b996-625cn" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.642087 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/878b0a76-9c37-4210-870b-5fed33721831-combined-ca-bundle\") pod \"placement-5cdcf9b996-625cn\" (UID: \"878b0a76-9c37-4210-870b-5fed33721831\") " pod="openstack/placement-5cdcf9b996-625cn" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.668424 4967 scope.go:117] "RemoveContainer" containerID="6b48779a5bbe12fa36091acda9cf3f4bca4678cb161b66dd0ed62b30620a5c7c" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.702228 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/keystone-76b6798d84-nx4fq"] Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.703886 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-76b6798d84-nx4fq" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.710265 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-config-data" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.710528 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-public-svc" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.710569 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-keystone-dockercfg-k7kxn" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.710652 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-76b6798d84-nx4fq"] Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.710777 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-keystone-internal-svc" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.710990 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone-scripts" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.712083 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"keystone" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.718227 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.734670 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.744261 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/878b0a76-9c37-4210-870b-5fed33721831-public-tls-certs\") pod \"placement-5cdcf9b996-625cn\" (UID: \"878b0a76-9c37-4210-870b-5fed33721831\") " pod="openstack/placement-5cdcf9b996-625cn" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.744298 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/878b0a76-9c37-4210-870b-5fed33721831-internal-tls-certs\") pod \"placement-5cdcf9b996-625cn\" (UID: \"878b0a76-9c37-4210-870b-5fed33721831\") " pod="openstack/placement-5cdcf9b996-625cn" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.744336 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6fhn\" (UniqueName: \"kubernetes.io/projected/878b0a76-9c37-4210-870b-5fed33721831-kube-api-access-s6fhn\") pod \"placement-5cdcf9b996-625cn\" (UID: \"878b0a76-9c37-4210-870b-5fed33721831\") " pod="openstack/placement-5cdcf9b996-625cn" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.744369 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/878b0a76-9c37-4210-870b-5fed33721831-config-data\") pod \"placement-5cdcf9b996-625cn\" (UID: \"878b0a76-9c37-4210-870b-5fed33721831\") " pod="openstack/placement-5cdcf9b996-625cn" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.744421 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/878b0a76-9c37-4210-870b-5fed33721831-logs\") pod \"placement-5cdcf9b996-625cn\" (UID: \"878b0a76-9c37-4210-870b-5fed33721831\") " pod="openstack/placement-5cdcf9b996-625cn" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.744449 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/878b0a76-9c37-4210-870b-5fed33721831-scripts\") pod \"placement-5cdcf9b996-625cn\" (UID: \"878b0a76-9c37-4210-870b-5fed33721831\") " pod="openstack/placement-5cdcf9b996-625cn" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.744494 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/878b0a76-9c37-4210-870b-5fed33721831-combined-ca-bundle\") pod \"placement-5cdcf9b996-625cn\" (UID: \"878b0a76-9c37-4210-870b-5fed33721831\") " pod="openstack/placement-5cdcf9b996-625cn" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.746064 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/878b0a76-9c37-4210-870b-5fed33721831-logs\") pod \"placement-5cdcf9b996-625cn\" (UID: \"878b0a76-9c37-4210-870b-5fed33721831\") " pod="openstack/placement-5cdcf9b996-625cn" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.749655 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/878b0a76-9c37-4210-870b-5fed33721831-public-tls-certs\") pod \"placement-5cdcf9b996-625cn\" (UID: \"878b0a76-9c37-4210-870b-5fed33721831\") " pod="openstack/placement-5cdcf9b996-625cn" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.751896 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/878b0a76-9c37-4210-870b-5fed33721831-internal-tls-certs\") pod \"placement-5cdcf9b996-625cn\" (UID: \"878b0a76-9c37-4210-870b-5fed33721831\") " pod="openstack/placement-5cdcf9b996-625cn" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.754386 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/878b0a76-9c37-4210-870b-5fed33721831-scripts\") pod \"placement-5cdcf9b996-625cn\" (UID: \"878b0a76-9c37-4210-870b-5fed33721831\") " pod="openstack/placement-5cdcf9b996-625cn" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.754435 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.755748 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/878b0a76-9c37-4210-870b-5fed33721831-config-data\") pod \"placement-5cdcf9b996-625cn\" (UID: \"878b0a76-9c37-4210-870b-5fed33721831\") " pod="openstack/placement-5cdcf9b996-625cn" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.757489 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.759967 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.760212 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.768446 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6fhn\" (UniqueName: \"kubernetes.io/projected/878b0a76-9c37-4210-870b-5fed33721831-kube-api-access-s6fhn\") pod \"placement-5cdcf9b996-625cn\" (UID: \"878b0a76-9c37-4210-870b-5fed33721831\") " pod="openstack/placement-5cdcf9b996-625cn" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.770197 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.774253 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/878b0a76-9c37-4210-870b-5fed33721831-combined-ca-bundle\") pod \"placement-5cdcf9b996-625cn\" (UID: \"878b0a76-9c37-4210-870b-5fed33721831\") " pod="openstack/placement-5cdcf9b996-625cn" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.845875 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb7b8d58-ee2e-4203-a1a9-44b2af5599cb-scripts\") pod \"keystone-76b6798d84-nx4fq\" (UID: \"eb7b8d58-ee2e-4203-a1a9-44b2af5599cb\") " pod="openstack/keystone-76b6798d84-nx4fq" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.845952 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eb7b8d58-ee2e-4203-a1a9-44b2af5599cb-fernet-keys\") pod \"keystone-76b6798d84-nx4fq\" (UID: \"eb7b8d58-ee2e-4203-a1a9-44b2af5599cb\") " pod="openstack/keystone-76b6798d84-nx4fq" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.845991 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb7b8d58-ee2e-4203-a1a9-44b2af5599cb-config-data\") pod \"keystone-76b6798d84-nx4fq\" (UID: \"eb7b8d58-ee2e-4203-a1a9-44b2af5599cb\") " pod="openstack/keystone-76b6798d84-nx4fq" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.846062 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb7b8d58-ee2e-4203-a1a9-44b2af5599cb-combined-ca-bundle\") pod \"keystone-76b6798d84-nx4fq\" (UID: \"eb7b8d58-ee2e-4203-a1a9-44b2af5599cb\") " pod="openstack/keystone-76b6798d84-nx4fq" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.846122 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svjkn\" (UniqueName: \"kubernetes.io/projected/eb7b8d58-ee2e-4203-a1a9-44b2af5599cb-kube-api-access-svjkn\") pod \"keystone-76b6798d84-nx4fq\" (UID: \"eb7b8d58-ee2e-4203-a1a9-44b2af5599cb\") " pod="openstack/keystone-76b6798d84-nx4fq" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.846177 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb7b8d58-ee2e-4203-a1a9-44b2af5599cb-public-tls-certs\") pod \"keystone-76b6798d84-nx4fq\" (UID: \"eb7b8d58-ee2e-4203-a1a9-44b2af5599cb\") " pod="openstack/keystone-76b6798d84-nx4fq" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.846209 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/eb7b8d58-ee2e-4203-a1a9-44b2af5599cb-credential-keys\") pod \"keystone-76b6798d84-nx4fq\" (UID: \"eb7b8d58-ee2e-4203-a1a9-44b2af5599cb\") " pod="openstack/keystone-76b6798d84-nx4fq" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.846257 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb7b8d58-ee2e-4203-a1a9-44b2af5599cb-internal-tls-certs\") pod \"keystone-76b6798d84-nx4fq\" (UID: \"eb7b8d58-ee2e-4203-a1a9-44b2af5599cb\") " pod="openstack/keystone-76b6798d84-nx4fq" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.948520 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb7b8d58-ee2e-4203-a1a9-44b2af5599cb-internal-tls-certs\") pod \"keystone-76b6798d84-nx4fq\" (UID: \"eb7b8d58-ee2e-4203-a1a9-44b2af5599cb\") " pod="openstack/keystone-76b6798d84-nx4fq" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.948600 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb7b8d58-ee2e-4203-a1a9-44b2af5599cb-scripts\") pod \"keystone-76b6798d84-nx4fq\" (UID: \"eb7b8d58-ee2e-4203-a1a9-44b2af5599cb\") " pod="openstack/keystone-76b6798d84-nx4fq" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.948631 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8df67d1-7a02-4e18-932a-433412376643-logs\") pod \"glance-default-internal-api-0\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.948651 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8df67d1-7a02-4e18-932a-433412376643-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.948682 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eb7b8d58-ee2e-4203-a1a9-44b2af5599cb-fernet-keys\") pod \"keystone-76b6798d84-nx4fq\" (UID: \"eb7b8d58-ee2e-4203-a1a9-44b2af5599cb\") " pod="openstack/keystone-76b6798d84-nx4fq" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.948727 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8df67d1-7a02-4e18-932a-433412376643-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.948745 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb7b8d58-ee2e-4203-a1a9-44b2af5599cb-config-data\") pod \"keystone-76b6798d84-nx4fq\" (UID: \"eb7b8d58-ee2e-4203-a1a9-44b2af5599cb\") " pod="openstack/keystone-76b6798d84-nx4fq" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.948788 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.948821 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8df67d1-7a02-4e18-932a-433412376643-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.948876 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb7b8d58-ee2e-4203-a1a9-44b2af5599cb-combined-ca-bundle\") pod \"keystone-76b6798d84-nx4fq\" (UID: \"eb7b8d58-ee2e-4203-a1a9-44b2af5599cb\") " pod="openstack/keystone-76b6798d84-nx4fq" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.948911 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8df67d1-7a02-4e18-932a-433412376643-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.948936 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d8df67d1-7a02-4e18-932a-433412376643-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.948985 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svjkn\" (UniqueName: \"kubernetes.io/projected/eb7b8d58-ee2e-4203-a1a9-44b2af5599cb-kube-api-access-svjkn\") pod \"keystone-76b6798d84-nx4fq\" (UID: \"eb7b8d58-ee2e-4203-a1a9-44b2af5599cb\") " pod="openstack/keystone-76b6798d84-nx4fq" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.949047 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m7jb4\" (UniqueName: \"kubernetes.io/projected/d8df67d1-7a02-4e18-932a-433412376643-kube-api-access-m7jb4\") pod \"glance-default-internal-api-0\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.949098 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb7b8d58-ee2e-4203-a1a9-44b2af5599cb-public-tls-certs\") pod \"keystone-76b6798d84-nx4fq\" (UID: \"eb7b8d58-ee2e-4203-a1a9-44b2af5599cb\") " pod="openstack/keystone-76b6798d84-nx4fq" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.949122 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/eb7b8d58-ee2e-4203-a1a9-44b2af5599cb-credential-keys\") pod \"keystone-76b6798d84-nx4fq\" (UID: \"eb7b8d58-ee2e-4203-a1a9-44b2af5599cb\") " pod="openstack/keystone-76b6798d84-nx4fq" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.955749 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"fernet-keys\" (UniqueName: \"kubernetes.io/secret/eb7b8d58-ee2e-4203-a1a9-44b2af5599cb-fernet-keys\") pod \"keystone-76b6798d84-nx4fq\" (UID: \"eb7b8d58-ee2e-4203-a1a9-44b2af5599cb\") " pod="openstack/keystone-76b6798d84-nx4fq" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.956176 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb7b8d58-ee2e-4203-a1a9-44b2af5599cb-scripts\") pod \"keystone-76b6798d84-nx4fq\" (UID: \"eb7b8d58-ee2e-4203-a1a9-44b2af5599cb\") " pod="openstack/keystone-76b6798d84-nx4fq" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.960018 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb7b8d58-ee2e-4203-a1a9-44b2af5599cb-public-tls-certs\") pod \"keystone-76b6798d84-nx4fq\" (UID: \"eb7b8d58-ee2e-4203-a1a9-44b2af5599cb\") " pod="openstack/keystone-76b6798d84-nx4fq" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.960347 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb7b8d58-ee2e-4203-a1a9-44b2af5599cb-config-data\") pod \"keystone-76b6798d84-nx4fq\" (UID: \"eb7b8d58-ee2e-4203-a1a9-44b2af5599cb\") " pod="openstack/keystone-76b6798d84-nx4fq" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.961592 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/eb7b8d58-ee2e-4203-a1a9-44b2af5599cb-internal-tls-certs\") pod \"keystone-76b6798d84-nx4fq\" (UID: \"eb7b8d58-ee2e-4203-a1a9-44b2af5599cb\") " pod="openstack/keystone-76b6798d84-nx4fq" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.967797 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb7b8d58-ee2e-4203-a1a9-44b2af5599cb-combined-ca-bundle\") pod \"keystone-76b6798d84-nx4fq\" (UID: \"eb7b8d58-ee2e-4203-a1a9-44b2af5599cb\") " pod="openstack/keystone-76b6798d84-nx4fq" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.968507 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"credential-keys\" (UniqueName: \"kubernetes.io/secret/eb7b8d58-ee2e-4203-a1a9-44b2af5599cb-credential-keys\") pod \"keystone-76b6798d84-nx4fq\" (UID: \"eb7b8d58-ee2e-4203-a1a9-44b2af5599cb\") " pod="openstack/keystone-76b6798d84-nx4fq" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.973015 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svjkn\" (UniqueName: \"kubernetes.io/projected/eb7b8d58-ee2e-4203-a1a9-44b2af5599cb-kube-api-access-svjkn\") pod \"keystone-76b6798d84-nx4fq\" (UID: \"eb7b8d58-ee2e-4203-a1a9-44b2af5599cb\") " pod="openstack/keystone-76b6798d84-nx4fq" Oct 11 04:07:47 crc kubenswrapper[4967]: I1011 04:07:47.994845 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/placement-5cdcf9b996-625cn" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.027010 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/keystone-76b6798d84-nx4fq" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.050266 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d8df67d1-7a02-4e18-932a-433412376643-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.050367 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m7jb4\" (UniqueName: \"kubernetes.io/projected/d8df67d1-7a02-4e18-932a-433412376643-kube-api-access-m7jb4\") pod \"glance-default-internal-api-0\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.050423 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8df67d1-7a02-4e18-932a-433412376643-logs\") pod \"glance-default-internal-api-0\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.050442 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8df67d1-7a02-4e18-932a-433412376643-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.050476 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8df67d1-7a02-4e18-932a-433412376643-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.050500 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.050521 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8df67d1-7a02-4e18-932a-433412376643-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.050553 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8df67d1-7a02-4e18-932a-433412376643-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.051283 4967 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.051584 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8df67d1-7a02-4e18-932a-433412376643-logs\") pod \"glance-default-internal-api-0\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.053265 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d8df67d1-7a02-4e18-932a-433412376643-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.055880 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8df67d1-7a02-4e18-932a-433412376643-scripts\") pod \"glance-default-internal-api-0\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.062295 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8df67d1-7a02-4e18-932a-433412376643-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.062417 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8df67d1-7a02-4e18-932a-433412376643-config-data\") pod \"glance-default-internal-api-0\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.073938 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8df67d1-7a02-4e18-932a-433412376643-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.077523 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m7jb4\" (UniqueName: \"kubernetes.io/projected/d8df67d1-7a02-4e18-932a-433412376643-kube-api-access-m7jb4\") pod \"glance-default-internal-api-0\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.114834 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.154259 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.594123 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-db-sync-sdg2s"] Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.595846 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-sync-sdg2s" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.597776 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-scripts" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.597952 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-ironic-dockercfg-vjmsm" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.597992 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-config-data" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.603947 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-db-sync-sdg2s"] Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.707966 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.708017 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.755932 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.762153 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.762498 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-75mhg\" (UniqueName: \"kubernetes.io/projected/1d249b36-580e-424d-8f36-e366c932411c-kube-api-access-75mhg\") pod \"ironic-db-sync-sdg2s\" (UID: \"1d249b36-580e-424d-8f36-e366c932411c\") " pod="openstack/ironic-db-sync-sdg2s" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.762692 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d249b36-580e-424d-8f36-e366c932411c-combined-ca-bundle\") pod \"ironic-db-sync-sdg2s\" (UID: \"1d249b36-580e-424d-8f36-e366c932411c\") " pod="openstack/ironic-db-sync-sdg2s" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.762757 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d249b36-580e-424d-8f36-e366c932411c-config-data\") pod \"ironic-db-sync-sdg2s\" (UID: \"1d249b36-580e-424d-8f36-e366c932411c\") " pod="openstack/ironic-db-sync-sdg2s" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.762873 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/1d249b36-580e-424d-8f36-e366c932411c-etc-podinfo\") pod \"ironic-db-sync-sdg2s\" (UID: \"1d249b36-580e-424d-8f36-e366c932411c\") " pod="openstack/ironic-db-sync-sdg2s" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.762911 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/1d249b36-580e-424d-8f36-e366c932411c-config-data-merged\") pod \"ironic-db-sync-sdg2s\" (UID: \"1d249b36-580e-424d-8f36-e366c932411c\") " pod="openstack/ironic-db-sync-sdg2s" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.762966 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d249b36-580e-424d-8f36-e366c932411c-scripts\") pod \"ironic-db-sync-sdg2s\" (UID: \"1d249b36-580e-424d-8f36-e366c932411c\") " pod="openstack/ironic-db-sync-sdg2s" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.831637 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5772c259-389a-4faf-8307-f14eff408981" path="/var/lib/kubelet/pods/5772c259-389a-4faf-8307-f14eff408981/volumes" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.865201 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d249b36-580e-424d-8f36-e366c932411c-combined-ca-bundle\") pod \"ironic-db-sync-sdg2s\" (UID: \"1d249b36-580e-424d-8f36-e366c932411c\") " pod="openstack/ironic-db-sync-sdg2s" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.865263 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d249b36-580e-424d-8f36-e366c932411c-config-data\") pod \"ironic-db-sync-sdg2s\" (UID: \"1d249b36-580e-424d-8f36-e366c932411c\") " pod="openstack/ironic-db-sync-sdg2s" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.865304 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/1d249b36-580e-424d-8f36-e366c932411c-etc-podinfo\") pod \"ironic-db-sync-sdg2s\" (UID: \"1d249b36-580e-424d-8f36-e366c932411c\") " pod="openstack/ironic-db-sync-sdg2s" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.865324 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/1d249b36-580e-424d-8f36-e366c932411c-config-data-merged\") pod \"ironic-db-sync-sdg2s\" (UID: \"1d249b36-580e-424d-8f36-e366c932411c\") " pod="openstack/ironic-db-sync-sdg2s" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.865347 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d249b36-580e-424d-8f36-e366c932411c-scripts\") pod \"ironic-db-sync-sdg2s\" (UID: \"1d249b36-580e-424d-8f36-e366c932411c\") " pod="openstack/ironic-db-sync-sdg2s" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.865406 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-75mhg\" (UniqueName: \"kubernetes.io/projected/1d249b36-580e-424d-8f36-e366c932411c-kube-api-access-75mhg\") pod \"ironic-db-sync-sdg2s\" (UID: \"1d249b36-580e-424d-8f36-e366c932411c\") " pod="openstack/ironic-db-sync-sdg2s" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.866429 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/1d249b36-580e-424d-8f36-e366c932411c-config-data-merged\") pod \"ironic-db-sync-sdg2s\" (UID: \"1d249b36-580e-424d-8f36-e366c932411c\") " pod="openstack/ironic-db-sync-sdg2s" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.869466 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d249b36-580e-424d-8f36-e366c932411c-combined-ca-bundle\") pod \"ironic-db-sync-sdg2s\" (UID: \"1d249b36-580e-424d-8f36-e366c932411c\") " pod="openstack/ironic-db-sync-sdg2s" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.871709 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d249b36-580e-424d-8f36-e366c932411c-config-data\") pod \"ironic-db-sync-sdg2s\" (UID: \"1d249b36-580e-424d-8f36-e366c932411c\") " pod="openstack/ironic-db-sync-sdg2s" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.882630 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d249b36-580e-424d-8f36-e366c932411c-scripts\") pod \"ironic-db-sync-sdg2s\" (UID: \"1d249b36-580e-424d-8f36-e366c932411c\") " pod="openstack/ironic-db-sync-sdg2s" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.884587 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/1d249b36-580e-424d-8f36-e366c932411c-etc-podinfo\") pod \"ironic-db-sync-sdg2s\" (UID: \"1d249b36-580e-424d-8f36-e366c932411c\") " pod="openstack/ironic-db-sync-sdg2s" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.886445 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-75mhg\" (UniqueName: \"kubernetes.io/projected/1d249b36-580e-424d-8f36-e366c932411c-kube-api-access-75mhg\") pod \"ironic-db-sync-sdg2s\" (UID: \"1d249b36-580e-424d-8f36-e366c932411c\") " pod="openstack/ironic-db-sync-sdg2s" Oct 11 04:07:48 crc kubenswrapper[4967]: I1011 04:07:48.918973 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-sync-sdg2s" Oct 11 04:07:49 crc kubenswrapper[4967]: I1011 04:07:49.656292 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 11 04:07:49 crc kubenswrapper[4967]: I1011 04:07:49.656707 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 11 04:07:51 crc kubenswrapper[4967]: I1011 04:07:51.572476 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 11 04:07:51 crc kubenswrapper[4967]: I1011 04:07:51.573966 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 11 04:07:52 crc kubenswrapper[4967]: I1011 04:07:52.527918 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 11 04:07:55 crc kubenswrapper[4967]: I1011 04:07:55.596875 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/placement-5cdcf9b996-625cn"] Oct 11 04:07:55 crc kubenswrapper[4967]: I1011 04:07:55.715009 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d8df67d1-7a02-4e18-932a-433412376643","Type":"ContainerStarted","Data":"4df1454b628b116376035dd87d97c74e4a376d13136837542258325283d9fb41"} Oct 11 04:08:01 crc kubenswrapper[4967]: I1011 04:08:01.769794 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5cdcf9b996-625cn" event={"ID":"878b0a76-9c37-4210-870b-5fed33721831","Type":"ContainerStarted","Data":"77142374c07a441278bce5c3e8af07cc94ff0d322e8bcbc087ab8fa2d7b25c45"} Oct 11 04:08:02 crc kubenswrapper[4967]: E1011 04:08:02.673690 4967 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying config: context canceled" image="quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified" Oct 11 04:08:02 crc kubenswrapper[4967]: E1011 04:08:02.674134 4967 kuberuntime_manager.go:1274] "Unhandled Error" err="container &Container{Name:cinder-db-sync,Image:quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified,Command:[/bin/bash],Args:[-c /usr/local/bin/kolla_set_configs && /usr/local/bin/kolla_start],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{EnvVar{Name:KOLLA_BOOTSTRAP,Value:TRUE,ValueFrom:nil,},EnvVar{Name:KOLLA_CONFIG_STRATEGY,Value:COPY_ALWAYS,ValueFrom:nil,},},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:etc-machine-id,ReadOnly:true,MountPath:/etc/machine-id,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:scripts,ReadOnly:true,MountPath:/usr/local/bin/container-scripts,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/config-data/merged,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/etc/my.cnf,SubPath:my.cnf,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:db-sync-config-data,ReadOnly:true,MountPath:/etc/cinder/cinder.conf.d,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:config-data,ReadOnly:true,MountPath:/var/lib/kolla/config_files/config.json,SubPath:db-sync-config.json,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:combined-ca-bundle,ReadOnly:true,MountPath:/etc/pki/ca-trust/extracted/pem/tls-ca-bundle.pem,SubPath:tls-ca-bundle.pem,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-dq4ts,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:IfNotPresent,SecurityContext:&SecurityContext{Capabilities:nil,Privileged:nil,SELinuxOptions:nil,RunAsUser:*0,RunAsNonRoot:nil,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:nil,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod cinder-db-sync-6bhhf_openstack(1680354a-b333-4e21-9748-0062d78a479c): ErrImagePull: rpc error: code = Canceled desc = copying config: context canceled" logger="UnhandledError" Oct 11 04:08:02 crc kubenswrapper[4967]: E1011 04:08:02.675297 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ErrImagePull: \"rpc error: code = Canceled desc = copying config: context canceled\"" pod="openstack/cinder-db-sync-6bhhf" podUID="1680354a-b333-4e21-9748-0062d78a479c" Oct 11 04:08:02 crc kubenswrapper[4967]: E1011 04:08:02.800072 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"cinder-db-sync\" with ImagePullBackOff: \"Back-off pulling image \\\"quay.io/podified-antelope-centos9/openstack-cinder-api:current-podified\\\"\"" pod="openstack/cinder-db-sync-6bhhf" podUID="1680354a-b333-4e21-9748-0062d78a479c" Oct 11 04:08:03 crc kubenswrapper[4967]: I1011 04:08:03.102846 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/keystone-76b6798d84-nx4fq"] Oct 11 04:08:03 crc kubenswrapper[4967]: I1011 04:08:03.182924 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-db-sync-sdg2s"] Oct 11 04:08:03 crc kubenswrapper[4967]: I1011 04:08:03.824567 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-sync-sdg2s" event={"ID":"1d249b36-580e-424d-8f36-e366c932411c","Type":"ContainerStarted","Data":"76a842317569015afefcd4c722bd5d59e42b98c49beaa46907bf45a29d5ca90d"} Oct 11 04:08:03 crc kubenswrapper[4967]: I1011 04:08:03.827330 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5cdcf9b996-625cn" event={"ID":"878b0a76-9c37-4210-870b-5fed33721831","Type":"ContainerStarted","Data":"9eb320c452976e0057b3e15f694d4e529d84ab6f6788677d29c9ccec6d188925"} Oct 11 04:08:03 crc kubenswrapper[4967]: I1011 04:08:03.827358 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/placement-5cdcf9b996-625cn" event={"ID":"878b0a76-9c37-4210-870b-5fed33721831","Type":"ContainerStarted","Data":"c6a9a777c581787327bfb53caa49c7ca9944823fe6389c4a322575b56ed89987"} Oct 11 04:08:03 crc kubenswrapper[4967]: I1011 04:08:03.827574 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5cdcf9b996-625cn" Oct 11 04:08:03 crc kubenswrapper[4967]: I1011 04:08:03.830793 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb56ccd5-8cd3-4da0-b11f-8600af8abf86","Type":"ContainerStarted","Data":"41b3bcad4054fd47b10fc41d843fdb5fbed7ef7a180a89424017ebb74d7b6fa7"} Oct 11 04:08:03 crc kubenswrapper[4967]: I1011 04:08:03.833139 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d8df67d1-7a02-4e18-932a-433412376643","Type":"ContainerStarted","Data":"78ddb9d89814b4e9e287119e168712ef35d876c9d62bc9a8ead0ec4db0ae1f75"} Oct 11 04:08:03 crc kubenswrapper[4967]: I1011 04:08:03.833193 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d8df67d1-7a02-4e18-932a-433412376643","Type":"ContainerStarted","Data":"223177e0c8f9782ee3f095403a0314a51431ed701e59885aa682f2114dd478b2"} Oct 11 04:08:03 crc kubenswrapper[4967]: I1011 04:08:03.835482 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-76b6798d84-nx4fq" event={"ID":"eb7b8d58-ee2e-4203-a1a9-44b2af5599cb","Type":"ContainerStarted","Data":"e19015f6f52e7f5021f5882f84081bdb6bdd075f4ef29e13979e50a2a497ccbf"} Oct 11 04:08:03 crc kubenswrapper[4967]: I1011 04:08:03.835551 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/keystone-76b6798d84-nx4fq" event={"ID":"eb7b8d58-ee2e-4203-a1a9-44b2af5599cb","Type":"ContainerStarted","Data":"550d2018da30f5939ec8309d49827b54c380e5bbf9bf4909a9e4f59c25283493"} Oct 11 04:08:03 crc kubenswrapper[4967]: I1011 04:08:03.835669 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/keystone-76b6798d84-nx4fq" Oct 11 04:08:03 crc kubenswrapper[4967]: I1011 04:08:03.836972 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-wdc9t" event={"ID":"8539e0a2-fecc-46f8-aa2f-a931a31cf17f","Type":"ContainerStarted","Data":"6fb4470936d71c0ec45dcea895c56f029bbb9d071929b0c549bb74cb2065a671"} Oct 11 04:08:03 crc kubenswrapper[4967]: I1011 04:08:03.854098 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/placement-5cdcf9b996-625cn" podStartSLOduration=16.854079281 podStartE2EDuration="16.854079281s" podCreationTimestamp="2025-10-11 04:07:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:08:03.848793249 +0000 UTC m=+1011.812002182" watchObservedRunningTime="2025-10-11 04:08:03.854079281 +0000 UTC m=+1011.817288204" Oct 11 04:08:03 crc kubenswrapper[4967]: I1011 04:08:03.870243 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-db-sync-wdc9t" podStartSLOduration=2.219385623 podStartE2EDuration="20.870226734s" podCreationTimestamp="2025-10-11 04:07:43 +0000 UTC" firstStartedPulling="2025-10-11 04:07:43.986484602 +0000 UTC m=+991.949693545" lastFinishedPulling="2025-10-11 04:08:02.637325723 +0000 UTC m=+1010.600534656" observedRunningTime="2025-10-11 04:08:03.866794524 +0000 UTC m=+1011.830003457" watchObservedRunningTime="2025-10-11 04:08:03.870226734 +0000 UTC m=+1011.833435667" Oct 11 04:08:03 crc kubenswrapper[4967]: I1011 04:08:03.883806 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/keystone-76b6798d84-nx4fq" podStartSLOduration=16.883786776 podStartE2EDuration="16.883786776s" podCreationTimestamp="2025-10-11 04:07:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:08:03.882274531 +0000 UTC m=+1011.845483464" watchObservedRunningTime="2025-10-11 04:08:03.883786776 +0000 UTC m=+1011.846995709" Oct 11 04:08:03 crc kubenswrapper[4967]: I1011 04:08:03.914337 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=16.9143133 podStartE2EDuration="16.9143133s" podCreationTimestamp="2025-10-11 04:07:47 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:08:03.903340777 +0000 UTC m=+1011.866549720" watchObservedRunningTime="2025-10-11 04:08:03.9143133 +0000 UTC m=+1011.877522253" Oct 11 04:08:04 crc kubenswrapper[4967]: I1011 04:08:04.851649 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/placement-5cdcf9b996-625cn" Oct 11 04:08:08 crc kubenswrapper[4967]: I1011 04:08:08.154941 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 11 04:08:08 crc kubenswrapper[4967]: I1011 04:08:08.157233 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 11 04:08:08 crc kubenswrapper[4967]: I1011 04:08:08.197721 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 11 04:08:08 crc kubenswrapper[4967]: I1011 04:08:08.203106 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 11 04:08:08 crc kubenswrapper[4967]: I1011 04:08:08.886953 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 11 04:08:08 crc kubenswrapper[4967]: I1011 04:08:08.887030 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 11 04:08:11 crc kubenswrapper[4967]: I1011 04:08:11.010263 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 11 04:08:11 crc kubenswrapper[4967]: I1011 04:08:11.010611 4967 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 11 04:08:11 crc kubenswrapper[4967]: I1011 04:08:11.012827 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 11 04:08:12 crc kubenswrapper[4967]: I1011 04:08:12.925228 4967 generic.go:334] "Generic (PLEG): container finished" podID="8539e0a2-fecc-46f8-aa2f-a931a31cf17f" containerID="6fb4470936d71c0ec45dcea895c56f029bbb9d071929b0c549bb74cb2065a671" exitCode=0 Oct 11 04:08:12 crc kubenswrapper[4967]: I1011 04:08:12.925322 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-wdc9t" event={"ID":"8539e0a2-fecc-46f8-aa2f-a931a31cf17f","Type":"ContainerDied","Data":"6fb4470936d71c0ec45dcea895c56f029bbb9d071929b0c549bb74cb2065a671"} Oct 11 04:08:13 crc kubenswrapper[4967]: I1011 04:08:13.935857 4967 generic.go:334] "Generic (PLEG): container finished" podID="28309ac7-8bd7-425c-a2f8-1041f1749acb" containerID="4f4319ea2749661e591abb39022c4bacf81f997235ba5a04b06afb53e61abb94" exitCode=0 Oct 11 04:08:13 crc kubenswrapper[4967]: I1011 04:08:13.936325 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-nqrkv" event={"ID":"28309ac7-8bd7-425c-a2f8-1041f1749acb","Type":"ContainerDied","Data":"4f4319ea2749661e591abb39022c4bacf81f997235ba5a04b06afb53e61abb94"} Oct 11 04:08:15 crc kubenswrapper[4967]: I1011 04:08:15.147811 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-wdc9t" Oct 11 04:08:15 crc kubenswrapper[4967]: I1011 04:08:15.274476 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8539e0a2-fecc-46f8-aa2f-a931a31cf17f-db-sync-config-data\") pod \"8539e0a2-fecc-46f8-aa2f-a931a31cf17f\" (UID: \"8539e0a2-fecc-46f8-aa2f-a931a31cf17f\") " Oct 11 04:08:15 crc kubenswrapper[4967]: I1011 04:08:15.274572 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-49m56\" (UniqueName: \"kubernetes.io/projected/8539e0a2-fecc-46f8-aa2f-a931a31cf17f-kube-api-access-49m56\") pod \"8539e0a2-fecc-46f8-aa2f-a931a31cf17f\" (UID: \"8539e0a2-fecc-46f8-aa2f-a931a31cf17f\") " Oct 11 04:08:15 crc kubenswrapper[4967]: I1011 04:08:15.274618 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8539e0a2-fecc-46f8-aa2f-a931a31cf17f-combined-ca-bundle\") pod \"8539e0a2-fecc-46f8-aa2f-a931a31cf17f\" (UID: \"8539e0a2-fecc-46f8-aa2f-a931a31cf17f\") " Oct 11 04:08:15 crc kubenswrapper[4967]: I1011 04:08:15.280023 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8539e0a2-fecc-46f8-aa2f-a931a31cf17f-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "8539e0a2-fecc-46f8-aa2f-a931a31cf17f" (UID: "8539e0a2-fecc-46f8-aa2f-a931a31cf17f"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:15 crc kubenswrapper[4967]: I1011 04:08:15.280541 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8539e0a2-fecc-46f8-aa2f-a931a31cf17f-kube-api-access-49m56" (OuterVolumeSpecName: "kube-api-access-49m56") pod "8539e0a2-fecc-46f8-aa2f-a931a31cf17f" (UID: "8539e0a2-fecc-46f8-aa2f-a931a31cf17f"). InnerVolumeSpecName "kube-api-access-49m56". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:08:15 crc kubenswrapper[4967]: I1011 04:08:15.305275 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8539e0a2-fecc-46f8-aa2f-a931a31cf17f-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8539e0a2-fecc-46f8-aa2f-a931a31cf17f" (UID: "8539e0a2-fecc-46f8-aa2f-a931a31cf17f"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:15 crc kubenswrapper[4967]: I1011 04:08:15.376628 4967 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/8539e0a2-fecc-46f8-aa2f-a931a31cf17f-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:15 crc kubenswrapper[4967]: I1011 04:08:15.376662 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-49m56\" (UniqueName: \"kubernetes.io/projected/8539e0a2-fecc-46f8-aa2f-a931a31cf17f-kube-api-access-49m56\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:15 crc kubenswrapper[4967]: I1011 04:08:15.376676 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8539e0a2-fecc-46f8-aa2f-a931a31cf17f-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:15 crc kubenswrapper[4967]: I1011 04:08:15.964019 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-db-sync-wdc9t" event={"ID":"8539e0a2-fecc-46f8-aa2f-a931a31cf17f","Type":"ContainerDied","Data":"3329e348c64cf3c0e690d218092c4f6adc2fbcfb3827fd51b3654641237bcf49"} Oct 11 04:08:15 crc kubenswrapper[4967]: I1011 04:08:15.964299 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="3329e348c64cf3c0e690d218092c4f6adc2fbcfb3827fd51b3654641237bcf49" Oct 11 04:08:15 crc kubenswrapper[4967]: I1011 04:08:15.964236 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-db-sync-wdc9t" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.438884 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-worker-6c87f65cff-p6j2x"] Oct 11 04:08:16 crc kubenswrapper[4967]: E1011 04:08:16.439227 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8539e0a2-fecc-46f8-aa2f-a931a31cf17f" containerName="barbican-db-sync" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.439241 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="8539e0a2-fecc-46f8-aa2f-a931a31cf17f" containerName="barbican-db-sync" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.439451 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="8539e0a2-fecc-46f8-aa2f-a931a31cf17f" containerName="barbican-db-sync" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.440433 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6c87f65cff-p6j2x" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.448829 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-worker-config-data" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.449168 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-barbican-dockercfg-mltgm" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.449365 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-config-data" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.454875 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-keystone-listener-7769b9cccd-zkh5v"] Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.457242 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7769b9cccd-zkh5v" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.459664 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-keystone-listener-config-data" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.499697 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ss5nv\" (UniqueName: \"kubernetes.io/projected/38e5cf89-c793-4207-b06e-d117da3dd152-kube-api-access-ss5nv\") pod \"barbican-keystone-listener-7769b9cccd-zkh5v\" (UID: \"38e5cf89-c793-4207-b06e-d117da3dd152\") " pod="openstack/barbican-keystone-listener-7769b9cccd-zkh5v" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.499739 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35-config-data\") pod \"barbican-worker-6c87f65cff-p6j2x\" (UID: \"fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35\") " pod="openstack/barbican-worker-6c87f65cff-p6j2x" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.499761 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38e5cf89-c793-4207-b06e-d117da3dd152-combined-ca-bundle\") pod \"barbican-keystone-listener-7769b9cccd-zkh5v\" (UID: \"38e5cf89-c793-4207-b06e-d117da3dd152\") " pod="openstack/barbican-keystone-listener-7769b9cccd-zkh5v" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.499782 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35-combined-ca-bundle\") pod \"barbican-worker-6c87f65cff-p6j2x\" (UID: \"fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35\") " pod="openstack/barbican-worker-6c87f65cff-p6j2x" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.499846 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35-logs\") pod \"barbican-worker-6c87f65cff-p6j2x\" (UID: \"fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35\") " pod="openstack/barbican-worker-6c87f65cff-p6j2x" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.499865 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35-config-data-custom\") pod \"barbican-worker-6c87f65cff-p6j2x\" (UID: \"fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35\") " pod="openstack/barbican-worker-6c87f65cff-p6j2x" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.499909 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38e5cf89-c793-4207-b06e-d117da3dd152-logs\") pod \"barbican-keystone-listener-7769b9cccd-zkh5v\" (UID: \"38e5cf89-c793-4207-b06e-d117da3dd152\") " pod="openstack/barbican-keystone-listener-7769b9cccd-zkh5v" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.499930 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h2gqx\" (UniqueName: \"kubernetes.io/projected/fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35-kube-api-access-h2gqx\") pod \"barbican-worker-6c87f65cff-p6j2x\" (UID: \"fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35\") " pod="openstack/barbican-worker-6c87f65cff-p6j2x" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.499956 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38e5cf89-c793-4207-b06e-d117da3dd152-config-data\") pod \"barbican-keystone-listener-7769b9cccd-zkh5v\" (UID: \"38e5cf89-c793-4207-b06e-d117da3dd152\") " pod="openstack/barbican-keystone-listener-7769b9cccd-zkh5v" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.499984 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/38e5cf89-c793-4207-b06e-d117da3dd152-config-data-custom\") pod \"barbican-keystone-listener-7769b9cccd-zkh5v\" (UID: \"38e5cf89-c793-4207-b06e-d117da3dd152\") " pod="openstack/barbican-keystone-listener-7769b9cccd-zkh5v" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.500338 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6c87f65cff-p6j2x"] Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.500878 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-nqrkv" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.509144 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7769b9cccd-zkh5v"] Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.566123 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-7c67bffd47-wgb47"] Oct 11 04:08:16 crc kubenswrapper[4967]: E1011 04:08:16.566880 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="28309ac7-8bd7-425c-a2f8-1041f1749acb" containerName="neutron-db-sync" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.566905 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="28309ac7-8bd7-425c-a2f8-1041f1749acb" containerName="neutron-db-sync" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.567111 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="28309ac7-8bd7-425c-a2f8-1041f1749acb" containerName="neutron-db-sync" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.567948 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c67bffd47-wgb47" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.579323 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c67bffd47-wgb47"] Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.600672 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d5j2j\" (UniqueName: \"kubernetes.io/projected/28309ac7-8bd7-425c-a2f8-1041f1749acb-kube-api-access-d5j2j\") pod \"28309ac7-8bd7-425c-a2f8-1041f1749acb\" (UID: \"28309ac7-8bd7-425c-a2f8-1041f1749acb\") " Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.600852 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28309ac7-8bd7-425c-a2f8-1041f1749acb-combined-ca-bundle\") pod \"28309ac7-8bd7-425c-a2f8-1041f1749acb\" (UID: \"28309ac7-8bd7-425c-a2f8-1041f1749acb\") " Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.600893 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/28309ac7-8bd7-425c-a2f8-1041f1749acb-config\") pod \"28309ac7-8bd7-425c-a2f8-1041f1749acb\" (UID: \"28309ac7-8bd7-425c-a2f8-1041f1749acb\") " Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.601115 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ss5nv\" (UniqueName: \"kubernetes.io/projected/38e5cf89-c793-4207-b06e-d117da3dd152-kube-api-access-ss5nv\") pod \"barbican-keystone-listener-7769b9cccd-zkh5v\" (UID: \"38e5cf89-c793-4207-b06e-d117da3dd152\") " pod="openstack/barbican-keystone-listener-7769b9cccd-zkh5v" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.601133 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35-config-data\") pod \"barbican-worker-6c87f65cff-p6j2x\" (UID: \"fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35\") " pod="openstack/barbican-worker-6c87f65cff-p6j2x" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.601152 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38e5cf89-c793-4207-b06e-d117da3dd152-combined-ca-bundle\") pod \"barbican-keystone-listener-7769b9cccd-zkh5v\" (UID: \"38e5cf89-c793-4207-b06e-d117da3dd152\") " pod="openstack/barbican-keystone-listener-7769b9cccd-zkh5v" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.601171 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3ae89d00-a652-471e-bea0-1ce8c690e257-dns-swift-storage-0\") pod \"dnsmasq-dns-7c67bffd47-wgb47\" (UID: \"3ae89d00-a652-471e-bea0-1ce8c690e257\") " pod="openstack/dnsmasq-dns-7c67bffd47-wgb47" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.601189 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35-combined-ca-bundle\") pod \"barbican-worker-6c87f65cff-p6j2x\" (UID: \"fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35\") " pod="openstack/barbican-worker-6c87f65cff-p6j2x" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.601217 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35-logs\") pod \"barbican-worker-6c87f65cff-p6j2x\" (UID: \"fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35\") " pod="openstack/barbican-worker-6c87f65cff-p6j2x" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.601237 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2x5dt\" (UniqueName: \"kubernetes.io/projected/3ae89d00-a652-471e-bea0-1ce8c690e257-kube-api-access-2x5dt\") pod \"dnsmasq-dns-7c67bffd47-wgb47\" (UID: \"3ae89d00-a652-471e-bea0-1ce8c690e257\") " pod="openstack/dnsmasq-dns-7c67bffd47-wgb47" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.601254 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35-config-data-custom\") pod \"barbican-worker-6c87f65cff-p6j2x\" (UID: \"fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35\") " pod="openstack/barbican-worker-6c87f65cff-p6j2x" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.601289 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ae89d00-a652-471e-bea0-1ce8c690e257-ovsdbserver-sb\") pod \"dnsmasq-dns-7c67bffd47-wgb47\" (UID: \"3ae89d00-a652-471e-bea0-1ce8c690e257\") " pod="openstack/dnsmasq-dns-7c67bffd47-wgb47" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.601313 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38e5cf89-c793-4207-b06e-d117da3dd152-logs\") pod \"barbican-keystone-listener-7769b9cccd-zkh5v\" (UID: \"38e5cf89-c793-4207-b06e-d117da3dd152\") " pod="openstack/barbican-keystone-listener-7769b9cccd-zkh5v" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.601335 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h2gqx\" (UniqueName: \"kubernetes.io/projected/fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35-kube-api-access-h2gqx\") pod \"barbican-worker-6c87f65cff-p6j2x\" (UID: \"fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35\") " pod="openstack/barbican-worker-6c87f65cff-p6j2x" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.601356 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ae89d00-a652-471e-bea0-1ce8c690e257-dns-svc\") pod \"dnsmasq-dns-7c67bffd47-wgb47\" (UID: \"3ae89d00-a652-471e-bea0-1ce8c690e257\") " pod="openstack/dnsmasq-dns-7c67bffd47-wgb47" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.601377 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38e5cf89-c793-4207-b06e-d117da3dd152-config-data\") pod \"barbican-keystone-listener-7769b9cccd-zkh5v\" (UID: \"38e5cf89-c793-4207-b06e-d117da3dd152\") " pod="openstack/barbican-keystone-listener-7769b9cccd-zkh5v" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.601399 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/38e5cf89-c793-4207-b06e-d117da3dd152-config-data-custom\") pod \"barbican-keystone-listener-7769b9cccd-zkh5v\" (UID: \"38e5cf89-c793-4207-b06e-d117da3dd152\") " pod="openstack/barbican-keystone-listener-7769b9cccd-zkh5v" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.601432 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ae89d00-a652-471e-bea0-1ce8c690e257-config\") pod \"dnsmasq-dns-7c67bffd47-wgb47\" (UID: \"3ae89d00-a652-471e-bea0-1ce8c690e257\") " pod="openstack/dnsmasq-dns-7c67bffd47-wgb47" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.601463 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ae89d00-a652-471e-bea0-1ce8c690e257-ovsdbserver-nb\") pod \"dnsmasq-dns-7c67bffd47-wgb47\" (UID: \"3ae89d00-a652-471e-bea0-1ce8c690e257\") " pod="openstack/dnsmasq-dns-7c67bffd47-wgb47" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.604198 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/38e5cf89-c793-4207-b06e-d117da3dd152-logs\") pod \"barbican-keystone-listener-7769b9cccd-zkh5v\" (UID: \"38e5cf89-c793-4207-b06e-d117da3dd152\") " pod="openstack/barbican-keystone-listener-7769b9cccd-zkh5v" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.605630 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/28309ac7-8bd7-425c-a2f8-1041f1749acb-kube-api-access-d5j2j" (OuterVolumeSpecName: "kube-api-access-d5j2j") pod "28309ac7-8bd7-425c-a2f8-1041f1749acb" (UID: "28309ac7-8bd7-425c-a2f8-1041f1749acb"). InnerVolumeSpecName "kube-api-access-d5j2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.608585 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35-logs\") pod \"barbican-worker-6c87f65cff-p6j2x\" (UID: \"fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35\") " pod="openstack/barbican-worker-6c87f65cff-p6j2x" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.629498 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/38e5cf89-c793-4207-b06e-d117da3dd152-config-data\") pod \"barbican-keystone-listener-7769b9cccd-zkh5v\" (UID: \"38e5cf89-c793-4207-b06e-d117da3dd152\") " pod="openstack/barbican-keystone-listener-7769b9cccd-zkh5v" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.630971 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35-combined-ca-bundle\") pod \"barbican-worker-6c87f65cff-p6j2x\" (UID: \"fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35\") " pod="openstack/barbican-worker-6c87f65cff-p6j2x" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.636101 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/38e5cf89-c793-4207-b06e-d117da3dd152-config-data-custom\") pod \"barbican-keystone-listener-7769b9cccd-zkh5v\" (UID: \"38e5cf89-c793-4207-b06e-d117da3dd152\") " pod="openstack/barbican-keystone-listener-7769b9cccd-zkh5v" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.640299 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h2gqx\" (UniqueName: \"kubernetes.io/projected/fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35-kube-api-access-h2gqx\") pod \"barbican-worker-6c87f65cff-p6j2x\" (UID: \"fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35\") " pod="openstack/barbican-worker-6c87f65cff-p6j2x" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.642301 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ss5nv\" (UniqueName: \"kubernetes.io/projected/38e5cf89-c793-4207-b06e-d117da3dd152-kube-api-access-ss5nv\") pod \"barbican-keystone-listener-7769b9cccd-zkh5v\" (UID: \"38e5cf89-c793-4207-b06e-d117da3dd152\") " pod="openstack/barbican-keystone-listener-7769b9cccd-zkh5v" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.642661 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/38e5cf89-c793-4207-b06e-d117da3dd152-combined-ca-bundle\") pod \"barbican-keystone-listener-7769b9cccd-zkh5v\" (UID: \"38e5cf89-c793-4207-b06e-d117da3dd152\") " pod="openstack/barbican-keystone-listener-7769b9cccd-zkh5v" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.643565 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35-config-data-custom\") pod \"barbican-worker-6c87f65cff-p6j2x\" (UID: \"fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35\") " pod="openstack/barbican-worker-6c87f65cff-p6j2x" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.645524 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35-config-data\") pod \"barbican-worker-6c87f65cff-p6j2x\" (UID: \"fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35\") " pod="openstack/barbican-worker-6c87f65cff-p6j2x" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.680895 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-6b6c6c8584-4qwkb"] Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.685868 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6b6c6c8584-4qwkb" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.691692 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"barbican-api-config-data" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.695172 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6b6c6c8584-4qwkb"] Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.696442 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28309ac7-8bd7-425c-a2f8-1041f1749acb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "28309ac7-8bd7-425c-a2f8-1041f1749acb" (UID: "28309ac7-8bd7-425c-a2f8-1041f1749acb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.705025 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9c6c522c-bd96-42f6-b0ae-27d55d3121c4-config-data-custom\") pod \"barbican-api-6b6c6c8584-4qwkb\" (UID: \"9c6c522c-bd96-42f6-b0ae-27d55d3121c4\") " pod="openstack/barbican-api-6b6c6c8584-4qwkb" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.705122 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2x5dt\" (UniqueName: \"kubernetes.io/projected/3ae89d00-a652-471e-bea0-1ce8c690e257-kube-api-access-2x5dt\") pod \"dnsmasq-dns-7c67bffd47-wgb47\" (UID: \"3ae89d00-a652-471e-bea0-1ce8c690e257\") " pod="openstack/dnsmasq-dns-7c67bffd47-wgb47" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.705181 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ae89d00-a652-471e-bea0-1ce8c690e257-ovsdbserver-sb\") pod \"dnsmasq-dns-7c67bffd47-wgb47\" (UID: \"3ae89d00-a652-471e-bea0-1ce8c690e257\") " pod="openstack/dnsmasq-dns-7c67bffd47-wgb47" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.705203 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c6c522c-bd96-42f6-b0ae-27d55d3121c4-logs\") pod \"barbican-api-6b6c6c8584-4qwkb\" (UID: \"9c6c522c-bd96-42f6-b0ae-27d55d3121c4\") " pod="openstack/barbican-api-6b6c6c8584-4qwkb" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.705226 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c6c522c-bd96-42f6-b0ae-27d55d3121c4-combined-ca-bundle\") pod \"barbican-api-6b6c6c8584-4qwkb\" (UID: \"9c6c522c-bd96-42f6-b0ae-27d55d3121c4\") " pod="openstack/barbican-api-6b6c6c8584-4qwkb" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.705250 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c6c522c-bd96-42f6-b0ae-27d55d3121c4-config-data\") pod \"barbican-api-6b6c6c8584-4qwkb\" (UID: \"9c6c522c-bd96-42f6-b0ae-27d55d3121c4\") " pod="openstack/barbican-api-6b6c6c8584-4qwkb" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.705273 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ae89d00-a652-471e-bea0-1ce8c690e257-dns-svc\") pod \"dnsmasq-dns-7c67bffd47-wgb47\" (UID: \"3ae89d00-a652-471e-bea0-1ce8c690e257\") " pod="openstack/dnsmasq-dns-7c67bffd47-wgb47" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.705343 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h95fg\" (UniqueName: \"kubernetes.io/projected/9c6c522c-bd96-42f6-b0ae-27d55d3121c4-kube-api-access-h95fg\") pod \"barbican-api-6b6c6c8584-4qwkb\" (UID: \"9c6c522c-bd96-42f6-b0ae-27d55d3121c4\") " pod="openstack/barbican-api-6b6c6c8584-4qwkb" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.705365 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ae89d00-a652-471e-bea0-1ce8c690e257-config\") pod \"dnsmasq-dns-7c67bffd47-wgb47\" (UID: \"3ae89d00-a652-471e-bea0-1ce8c690e257\") " pod="openstack/dnsmasq-dns-7c67bffd47-wgb47" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.705397 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ae89d00-a652-471e-bea0-1ce8c690e257-ovsdbserver-nb\") pod \"dnsmasq-dns-7c67bffd47-wgb47\" (UID: \"3ae89d00-a652-471e-bea0-1ce8c690e257\") " pod="openstack/dnsmasq-dns-7c67bffd47-wgb47" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.705453 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3ae89d00-a652-471e-bea0-1ce8c690e257-dns-swift-storage-0\") pod \"dnsmasq-dns-7c67bffd47-wgb47\" (UID: \"3ae89d00-a652-471e-bea0-1ce8c690e257\") " pod="openstack/dnsmasq-dns-7c67bffd47-wgb47" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.705494 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/28309ac7-8bd7-425c-a2f8-1041f1749acb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.705505 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d5j2j\" (UniqueName: \"kubernetes.io/projected/28309ac7-8bd7-425c-a2f8-1041f1749acb-kube-api-access-d5j2j\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.706626 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3ae89d00-a652-471e-bea0-1ce8c690e257-dns-swift-storage-0\") pod \"dnsmasq-dns-7c67bffd47-wgb47\" (UID: \"3ae89d00-a652-471e-bea0-1ce8c690e257\") " pod="openstack/dnsmasq-dns-7c67bffd47-wgb47" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.706999 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ae89d00-a652-471e-bea0-1ce8c690e257-dns-svc\") pod \"dnsmasq-dns-7c67bffd47-wgb47\" (UID: \"3ae89d00-a652-471e-bea0-1ce8c690e257\") " pod="openstack/dnsmasq-dns-7c67bffd47-wgb47" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.707851 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ae89d00-a652-471e-bea0-1ce8c690e257-ovsdbserver-sb\") pod \"dnsmasq-dns-7c67bffd47-wgb47\" (UID: \"3ae89d00-a652-471e-bea0-1ce8c690e257\") " pod="openstack/dnsmasq-dns-7c67bffd47-wgb47" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.708161 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ae89d00-a652-471e-bea0-1ce8c690e257-config\") pod \"dnsmasq-dns-7c67bffd47-wgb47\" (UID: \"3ae89d00-a652-471e-bea0-1ce8c690e257\") " pod="openstack/dnsmasq-dns-7c67bffd47-wgb47" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.708651 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ae89d00-a652-471e-bea0-1ce8c690e257-ovsdbserver-nb\") pod \"dnsmasq-dns-7c67bffd47-wgb47\" (UID: \"3ae89d00-a652-471e-bea0-1ce8c690e257\") " pod="openstack/dnsmasq-dns-7c67bffd47-wgb47" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.719346 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/28309ac7-8bd7-425c-a2f8-1041f1749acb-config" (OuterVolumeSpecName: "config") pod "28309ac7-8bd7-425c-a2f8-1041f1749acb" (UID: "28309ac7-8bd7-425c-a2f8-1041f1749acb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.725500 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2x5dt\" (UniqueName: \"kubernetes.io/projected/3ae89d00-a652-471e-bea0-1ce8c690e257-kube-api-access-2x5dt\") pod \"dnsmasq-dns-7c67bffd47-wgb47\" (UID: \"3ae89d00-a652-471e-bea0-1ce8c690e257\") " pod="openstack/dnsmasq-dns-7c67bffd47-wgb47" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.807146 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c6c522c-bd96-42f6-b0ae-27d55d3121c4-logs\") pod \"barbican-api-6b6c6c8584-4qwkb\" (UID: \"9c6c522c-bd96-42f6-b0ae-27d55d3121c4\") " pod="openstack/barbican-api-6b6c6c8584-4qwkb" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.807213 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c6c522c-bd96-42f6-b0ae-27d55d3121c4-combined-ca-bundle\") pod \"barbican-api-6b6c6c8584-4qwkb\" (UID: \"9c6c522c-bd96-42f6-b0ae-27d55d3121c4\") " pod="openstack/barbican-api-6b6c6c8584-4qwkb" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.807246 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c6c522c-bd96-42f6-b0ae-27d55d3121c4-config-data\") pod \"barbican-api-6b6c6c8584-4qwkb\" (UID: \"9c6c522c-bd96-42f6-b0ae-27d55d3121c4\") " pod="openstack/barbican-api-6b6c6c8584-4qwkb" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.807302 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h95fg\" (UniqueName: \"kubernetes.io/projected/9c6c522c-bd96-42f6-b0ae-27d55d3121c4-kube-api-access-h95fg\") pod \"barbican-api-6b6c6c8584-4qwkb\" (UID: \"9c6c522c-bd96-42f6-b0ae-27d55d3121c4\") " pod="openstack/barbican-api-6b6c6c8584-4qwkb" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.807390 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9c6c522c-bd96-42f6-b0ae-27d55d3121c4-config-data-custom\") pod \"barbican-api-6b6c6c8584-4qwkb\" (UID: \"9c6c522c-bd96-42f6-b0ae-27d55d3121c4\") " pod="openstack/barbican-api-6b6c6c8584-4qwkb" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.808053 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/28309ac7-8bd7-425c-a2f8-1041f1749acb-config\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.808271 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c6c522c-bd96-42f6-b0ae-27d55d3121c4-logs\") pod \"barbican-api-6b6c6c8584-4qwkb\" (UID: \"9c6c522c-bd96-42f6-b0ae-27d55d3121c4\") " pod="openstack/barbican-api-6b6c6c8584-4qwkb" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.811472 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c6c522c-bd96-42f6-b0ae-27d55d3121c4-combined-ca-bundle\") pod \"barbican-api-6b6c6c8584-4qwkb\" (UID: \"9c6c522c-bd96-42f6-b0ae-27d55d3121c4\") " pod="openstack/barbican-api-6b6c6c8584-4qwkb" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.811749 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9c6c522c-bd96-42f6-b0ae-27d55d3121c4-config-data-custom\") pod \"barbican-api-6b6c6c8584-4qwkb\" (UID: \"9c6c522c-bd96-42f6-b0ae-27d55d3121c4\") " pod="openstack/barbican-api-6b6c6c8584-4qwkb" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.814574 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c6c522c-bd96-42f6-b0ae-27d55d3121c4-config-data\") pod \"barbican-api-6b6c6c8584-4qwkb\" (UID: \"9c6c522c-bd96-42f6-b0ae-27d55d3121c4\") " pod="openstack/barbican-api-6b6c6c8584-4qwkb" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.826168 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h95fg\" (UniqueName: \"kubernetes.io/projected/9c6c522c-bd96-42f6-b0ae-27d55d3121c4-kube-api-access-h95fg\") pod \"barbican-api-6b6c6c8584-4qwkb\" (UID: \"9c6c522c-bd96-42f6-b0ae-27d55d3121c4\") " pod="openstack/barbican-api-6b6c6c8584-4qwkb" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.891960 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-worker-6c87f65cff-p6j2x" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.947584 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-keystone-listener-7769b9cccd-zkh5v" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.972133 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c67bffd47-wgb47" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.977912 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-db-sync-nqrkv" event={"ID":"28309ac7-8bd7-425c-a2f8-1041f1749acb","Type":"ContainerDied","Data":"b62926e006d4a5758e50d894276bb514a52435eebbc618847cae44df89e55bd9"} Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.977950 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="b62926e006d4a5758e50d894276bb514a52435eebbc618847cae44df89e55bd9" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.978030 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-db-sync-nqrkv" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.983510 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb56ccd5-8cd3-4da0-b11f-8600af8abf86","Type":"ContainerStarted","Data":"b73b0ac6958bcde79cb6d728ac0977bd7a8319a230cd00a4e2692a9e90ab183c"} Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.984084 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eb56ccd5-8cd3-4da0-b11f-8600af8abf86" containerName="ceilometer-central-agent" containerID="cri-o://26117f8c594b515e35c7c2108fef08a9638b8113dcdf4be3945b880c53864c0d" gracePeriod=30 Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.984207 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.984263 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eb56ccd5-8cd3-4da0-b11f-8600af8abf86" containerName="proxy-httpd" containerID="cri-o://b73b0ac6958bcde79cb6d728ac0977bd7a8319a230cd00a4e2692a9e90ab183c" gracePeriod=30 Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.984311 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eb56ccd5-8cd3-4da0-b11f-8600af8abf86" containerName="sg-core" containerID="cri-o://41b3bcad4054fd47b10fc41d843fdb5fbed7ef7a180a89424017ebb74d7b6fa7" gracePeriod=30 Oct 11 04:08:16 crc kubenswrapper[4967]: I1011 04:08:16.984357 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eb56ccd5-8cd3-4da0-b11f-8600af8abf86" containerName="ceilometer-notification-agent" containerID="cri-o://ec8b91f055b4e4c15ad8ebb839b6d61ae69225f0fd74f6083d4bc346cf2ab43b" gracePeriod=30 Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.009486 4967 generic.go:334] "Generic (PLEG): container finished" podID="1d249b36-580e-424d-8f36-e366c932411c" containerID="f81d31cd5813da20384292b38586fc79d502fb93f0d6acdd4be071ddc15c5ff9" exitCode=0 Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.009538 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-sync-sdg2s" event={"ID":"1d249b36-580e-424d-8f36-e366c932411c","Type":"ContainerDied","Data":"f81d31cd5813da20384292b38586fc79d502fb93f0d6acdd4be071ddc15c5ff9"} Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.042056 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6b6c6c8584-4qwkb" Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.060418 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.355498652 podStartE2EDuration="49.06039834s" podCreationTimestamp="2025-10-11 04:07:28 +0000 UTC" firstStartedPulling="2025-10-11 04:07:29.646712497 +0000 UTC m=+977.609921430" lastFinishedPulling="2025-10-11 04:08:16.351612185 +0000 UTC m=+1024.314821118" observedRunningTime="2025-10-11 04:08:17.004186503 +0000 UTC m=+1024.967395436" watchObservedRunningTime="2025-10-11 04:08:17.06039834 +0000 UTC m=+1025.023607273" Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.423042 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-worker-6c87f65cff-p6j2x"] Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.552792 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-keystone-listener-7769b9cccd-zkh5v"] Oct 11 04:08:17 crc kubenswrapper[4967]: W1011 04:08:17.559698 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod38e5cf89_c793_4207_b06e_d117da3dd152.slice/crio-ce51f167a131f2d12045df05459e0876fcd6c2e4226dd28f010d1f5ad84840fe WatchSource:0}: Error finding container ce51f167a131f2d12045df05459e0876fcd6c2e4226dd28f010d1f5ad84840fe: Status 404 returned error can't find the container with id ce51f167a131f2d12045df05459e0876fcd6c2e4226dd28f010d1f5ad84840fe Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.674541 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-7c67bffd47-wgb47"] Oct 11 04:08:17 crc kubenswrapper[4967]: W1011 04:08:17.686255 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3ae89d00_a652_471e_bea0_1ce8c690e257.slice/crio-435831d7c0933f1b223580056ad5882aa12216f78da38658b2772d88d12921a3 WatchSource:0}: Error finding container 435831d7c0933f1b223580056ad5882aa12216f78da38658b2772d88d12921a3: Status 404 returned error can't find the container with id 435831d7c0933f1b223580056ad5882aa12216f78da38658b2772d88d12921a3 Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.735582 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-6b6c6c8584-4qwkb"] Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.775275 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c67bffd47-wgb47"] Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.794611 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-jdrjp"] Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.796671 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-jdrjp" Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.824201 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-jdrjp"] Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.883175 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-jdrjp\" (UID: \"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43\") " pod="openstack/dnsmasq-dns-848cf88cfc-jdrjp" Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.883245 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-jdrjp\" (UID: \"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43\") " pod="openstack/dnsmasq-dns-848cf88cfc-jdrjp" Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.883299 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hrx7l\" (UniqueName: \"kubernetes.io/projected/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-kube-api-access-hrx7l\") pod \"dnsmasq-dns-848cf88cfc-jdrjp\" (UID: \"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43\") " pod="openstack/dnsmasq-dns-848cf88cfc-jdrjp" Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.883332 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-jdrjp\" (UID: \"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43\") " pod="openstack/dnsmasq-dns-848cf88cfc-jdrjp" Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.883360 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-jdrjp\" (UID: \"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43\") " pod="openstack/dnsmasq-dns-848cf88cfc-jdrjp" Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.883424 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-config\") pod \"dnsmasq-dns-848cf88cfc-jdrjp\" (UID: \"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43\") " pod="openstack/dnsmasq-dns-848cf88cfc-jdrjp" Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.887958 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-6554fc45b4-x7btb"] Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.889743 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6554fc45b4-x7btb" Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.894248 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-httpd-config" Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.894418 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-config" Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.894805 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-ovndbs" Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.894843 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"neutron-neutron-dockercfg-kpsm7" Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.901151 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6554fc45b4-x7btb"] Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.986274 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/92205841-5c2e-4fc7-9328-8a164f5fb8cf-config\") pod \"neutron-6554fc45b4-x7btb\" (UID: \"92205841-5c2e-4fc7-9328-8a164f5fb8cf\") " pod="openstack/neutron-6554fc45b4-x7btb" Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.986320 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hrx7l\" (UniqueName: \"kubernetes.io/projected/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-kube-api-access-hrx7l\") pod \"dnsmasq-dns-848cf88cfc-jdrjp\" (UID: \"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43\") " pod="openstack/dnsmasq-dns-848cf88cfc-jdrjp" Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.986357 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-jdrjp\" (UID: \"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43\") " pod="openstack/dnsmasq-dns-848cf88cfc-jdrjp" Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.986380 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-jdrjp\" (UID: \"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43\") " pod="openstack/dnsmasq-dns-848cf88cfc-jdrjp" Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.986433 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-config\") pod \"dnsmasq-dns-848cf88cfc-jdrjp\" (UID: \"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43\") " pod="openstack/dnsmasq-dns-848cf88cfc-jdrjp" Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.986475 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92205841-5c2e-4fc7-9328-8a164f5fb8cf-combined-ca-bundle\") pod \"neutron-6554fc45b4-x7btb\" (UID: \"92205841-5c2e-4fc7-9328-8a164f5fb8cf\") " pod="openstack/neutron-6554fc45b4-x7btb" Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.986500 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/92205841-5c2e-4fc7-9328-8a164f5fb8cf-httpd-config\") pod \"neutron-6554fc45b4-x7btb\" (UID: \"92205841-5c2e-4fc7-9328-8a164f5fb8cf\") " pod="openstack/neutron-6554fc45b4-x7btb" Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.986545 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/92205841-5c2e-4fc7-9328-8a164f5fb8cf-ovndb-tls-certs\") pod \"neutron-6554fc45b4-x7btb\" (UID: \"92205841-5c2e-4fc7-9328-8a164f5fb8cf\") " pod="openstack/neutron-6554fc45b4-x7btb" Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.986572 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ld8gb\" (UniqueName: \"kubernetes.io/projected/92205841-5c2e-4fc7-9328-8a164f5fb8cf-kube-api-access-ld8gb\") pod \"neutron-6554fc45b4-x7btb\" (UID: \"92205841-5c2e-4fc7-9328-8a164f5fb8cf\") " pod="openstack/neutron-6554fc45b4-x7btb" Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.986614 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-jdrjp\" (UID: \"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43\") " pod="openstack/dnsmasq-dns-848cf88cfc-jdrjp" Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.986656 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-jdrjp\" (UID: \"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43\") " pod="openstack/dnsmasq-dns-848cf88cfc-jdrjp" Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.987621 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-dns-swift-storage-0\") pod \"dnsmasq-dns-848cf88cfc-jdrjp\" (UID: \"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43\") " pod="openstack/dnsmasq-dns-848cf88cfc-jdrjp" Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.988602 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-dns-svc\") pod \"dnsmasq-dns-848cf88cfc-jdrjp\" (UID: \"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43\") " pod="openstack/dnsmasq-dns-848cf88cfc-jdrjp" Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.989176 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-ovsdbserver-nb\") pod \"dnsmasq-dns-848cf88cfc-jdrjp\" (UID: \"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43\") " pod="openstack/dnsmasq-dns-848cf88cfc-jdrjp" Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.989730 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-config\") pod \"dnsmasq-dns-848cf88cfc-jdrjp\" (UID: \"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43\") " pod="openstack/dnsmasq-dns-848cf88cfc-jdrjp" Oct 11 04:08:17 crc kubenswrapper[4967]: I1011 04:08:17.990182 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-ovsdbserver-sb\") pod \"dnsmasq-dns-848cf88cfc-jdrjp\" (UID: \"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43\") " pod="openstack/dnsmasq-dns-848cf88cfc-jdrjp" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.008771 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hrx7l\" (UniqueName: \"kubernetes.io/projected/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-kube-api-access-hrx7l\") pod \"dnsmasq-dns-848cf88cfc-jdrjp\" (UID: \"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43\") " pod="openstack/dnsmasq-dns-848cf88cfc-jdrjp" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.049459 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-6bhhf" event={"ID":"1680354a-b333-4e21-9748-0062d78a479c","Type":"ContainerStarted","Data":"433cef11bc0a2150c9ca5cf11665083f8de67f71d9f8a86c58ef07e0b0409f7e"} Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.067747 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-sync-sdg2s" event={"ID":"1d249b36-580e-424d-8f36-e366c932411c","Type":"ContainerStarted","Data":"02bf66457fc098fe15c42a14af57eb113a56800408703ed35ed3a2b4f07ab74e"} Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.075905 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-db-sync-6bhhf" podStartSLOduration=3.551421409 podStartE2EDuration="36.075883397s" podCreationTimestamp="2025-10-11 04:07:42 +0000 UTC" firstStartedPulling="2025-10-11 04:07:43.863553997 +0000 UTC m=+991.826762930" lastFinishedPulling="2025-10-11 04:08:16.388015975 +0000 UTC m=+1024.351224918" observedRunningTime="2025-10-11 04:08:18.067528374 +0000 UTC m=+1026.030737307" watchObservedRunningTime="2025-10-11 04:08:18.075883397 +0000 UTC m=+1026.039092330" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.080841 4967 generic.go:334] "Generic (PLEG): container finished" podID="3ae89d00-a652-471e-bea0-1ce8c690e257" containerID="809c6e9eb61c4ab3be8c3e690ce8d39887b347035df7db22bfbfff76fdc1fe03" exitCode=0 Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.080910 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c67bffd47-wgb47" event={"ID":"3ae89d00-a652-471e-bea0-1ce8c690e257","Type":"ContainerDied","Data":"809c6e9eb61c4ab3be8c3e690ce8d39887b347035df7db22bfbfff76fdc1fe03"} Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.080936 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c67bffd47-wgb47" event={"ID":"3ae89d00-a652-471e-bea0-1ce8c690e257","Type":"ContainerStarted","Data":"435831d7c0933f1b223580056ad5882aa12216f78da38658b2772d88d12921a3"} Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.084231 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6c87f65cff-p6j2x" event={"ID":"fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35","Type":"ContainerStarted","Data":"6c4161178553e60146550b687ff4944fe6bf2064c5d1989b2559aa7c9804014f"} Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.091433 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/92205841-5c2e-4fc7-9328-8a164f5fb8cf-ovndb-tls-certs\") pod \"neutron-6554fc45b4-x7btb\" (UID: \"92205841-5c2e-4fc7-9328-8a164f5fb8cf\") " pod="openstack/neutron-6554fc45b4-x7btb" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.091491 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ld8gb\" (UniqueName: \"kubernetes.io/projected/92205841-5c2e-4fc7-9328-8a164f5fb8cf-kube-api-access-ld8gb\") pod \"neutron-6554fc45b4-x7btb\" (UID: \"92205841-5c2e-4fc7-9328-8a164f5fb8cf\") " pod="openstack/neutron-6554fc45b4-x7btb" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.091502 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7769b9cccd-zkh5v" event={"ID":"38e5cf89-c793-4207-b06e-d117da3dd152","Type":"ContainerStarted","Data":"ce51f167a131f2d12045df05459e0876fcd6c2e4226dd28f010d1f5ad84840fe"} Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.091592 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/92205841-5c2e-4fc7-9328-8a164f5fb8cf-config\") pod \"neutron-6554fc45b4-x7btb\" (UID: \"92205841-5c2e-4fc7-9328-8a164f5fb8cf\") " pod="openstack/neutron-6554fc45b4-x7btb" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.091896 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92205841-5c2e-4fc7-9328-8a164f5fb8cf-combined-ca-bundle\") pod \"neutron-6554fc45b4-x7btb\" (UID: \"92205841-5c2e-4fc7-9328-8a164f5fb8cf\") " pod="openstack/neutron-6554fc45b4-x7btb" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.091933 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/92205841-5c2e-4fc7-9328-8a164f5fb8cf-httpd-config\") pod \"neutron-6554fc45b4-x7btb\" (UID: \"92205841-5c2e-4fc7-9328-8a164f5fb8cf\") " pod="openstack/neutron-6554fc45b4-x7btb" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.098261 4967 generic.go:334] "Generic (PLEG): container finished" podID="eb56ccd5-8cd3-4da0-b11f-8600af8abf86" containerID="b73b0ac6958bcde79cb6d728ac0977bd7a8319a230cd00a4e2692a9e90ab183c" exitCode=0 Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.098290 4967 generic.go:334] "Generic (PLEG): container finished" podID="eb56ccd5-8cd3-4da0-b11f-8600af8abf86" containerID="41b3bcad4054fd47b10fc41d843fdb5fbed7ef7a180a89424017ebb74d7b6fa7" exitCode=2 Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.098299 4967 generic.go:334] "Generic (PLEG): container finished" podID="eb56ccd5-8cd3-4da0-b11f-8600af8abf86" containerID="26117f8c594b515e35c7c2108fef08a9638b8113dcdf4be3945b880c53864c0d" exitCode=0 Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.098340 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb56ccd5-8cd3-4da0-b11f-8600af8abf86","Type":"ContainerDied","Data":"b73b0ac6958bcde79cb6d728ac0977bd7a8319a230cd00a4e2692a9e90ab183c"} Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.098364 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb56ccd5-8cd3-4da0-b11f-8600af8abf86","Type":"ContainerDied","Data":"41b3bcad4054fd47b10fc41d843fdb5fbed7ef7a180a89424017ebb74d7b6fa7"} Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.098375 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb56ccd5-8cd3-4da0-b11f-8600af8abf86","Type":"ContainerDied","Data":"26117f8c594b515e35c7c2108fef08a9638b8113dcdf4be3945b880c53864c0d"} Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.099510 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b6c6c8584-4qwkb" event={"ID":"9c6c522c-bd96-42f6-b0ae-27d55d3121c4","Type":"ContainerStarted","Data":"2713f1c59d64482f421e2cd9da89a876d483393a3a347f05ed054178dad43dfa"} Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.099538 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b6c6c8584-4qwkb" event={"ID":"9c6c522c-bd96-42f6-b0ae-27d55d3121c4","Type":"ContainerStarted","Data":"782fd2c4860269e8a59f9b8daf4a6230aa99ba0245156609df6fa27e88f34205"} Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.103574 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-db-sync-sdg2s" podStartSLOduration=16.904652337 podStartE2EDuration="30.103561015s" podCreationTimestamp="2025-10-11 04:07:48 +0000 UTC" firstStartedPulling="2025-10-11 04:08:03.187877818 +0000 UTC m=+1011.151086751" lastFinishedPulling="2025-10-11 04:08:16.386786496 +0000 UTC m=+1024.349995429" observedRunningTime="2025-10-11 04:08:18.090089615 +0000 UTC m=+1026.053298548" watchObservedRunningTime="2025-10-11 04:08:18.103561015 +0000 UTC m=+1026.066769948" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.104053 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92205841-5c2e-4fc7-9328-8a164f5fb8cf-combined-ca-bundle\") pod \"neutron-6554fc45b4-x7btb\" (UID: \"92205841-5c2e-4fc7-9328-8a164f5fb8cf\") " pod="openstack/neutron-6554fc45b4-x7btb" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.108957 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/92205841-5c2e-4fc7-9328-8a164f5fb8cf-ovndb-tls-certs\") pod \"neutron-6554fc45b4-x7btb\" (UID: \"92205841-5c2e-4fc7-9328-8a164f5fb8cf\") " pod="openstack/neutron-6554fc45b4-x7btb" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.113694 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ld8gb\" (UniqueName: \"kubernetes.io/projected/92205841-5c2e-4fc7-9328-8a164f5fb8cf-kube-api-access-ld8gb\") pod \"neutron-6554fc45b4-x7btb\" (UID: \"92205841-5c2e-4fc7-9328-8a164f5fb8cf\") " pod="openstack/neutron-6554fc45b4-x7btb" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.113940 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/92205841-5c2e-4fc7-9328-8a164f5fb8cf-httpd-config\") pod \"neutron-6554fc45b4-x7btb\" (UID: \"92205841-5c2e-4fc7-9328-8a164f5fb8cf\") " pod="openstack/neutron-6554fc45b4-x7btb" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.117460 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/92205841-5c2e-4fc7-9328-8a164f5fb8cf-config\") pod \"neutron-6554fc45b4-x7btb\" (UID: \"92205841-5c2e-4fc7-9328-8a164f5fb8cf\") " pod="openstack/neutron-6554fc45b4-x7btb" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.142758 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-jdrjp" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.230204 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6554fc45b4-x7btb" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.581484 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c67bffd47-wgb47" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.731577 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ae89d00-a652-471e-bea0-1ce8c690e257-config\") pod \"3ae89d00-a652-471e-bea0-1ce8c690e257\" (UID: \"3ae89d00-a652-471e-bea0-1ce8c690e257\") " Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.731983 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ae89d00-a652-471e-bea0-1ce8c690e257-ovsdbserver-nb\") pod \"3ae89d00-a652-471e-bea0-1ce8c690e257\" (UID: \"3ae89d00-a652-471e-bea0-1ce8c690e257\") " Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.732125 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2x5dt\" (UniqueName: \"kubernetes.io/projected/3ae89d00-a652-471e-bea0-1ce8c690e257-kube-api-access-2x5dt\") pod \"3ae89d00-a652-471e-bea0-1ce8c690e257\" (UID: \"3ae89d00-a652-471e-bea0-1ce8c690e257\") " Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.732149 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ae89d00-a652-471e-bea0-1ce8c690e257-ovsdbserver-sb\") pod \"3ae89d00-a652-471e-bea0-1ce8c690e257\" (UID: \"3ae89d00-a652-471e-bea0-1ce8c690e257\") " Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.732196 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3ae89d00-a652-471e-bea0-1ce8c690e257-dns-swift-storage-0\") pod \"3ae89d00-a652-471e-bea0-1ce8c690e257\" (UID: \"3ae89d00-a652-471e-bea0-1ce8c690e257\") " Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.732242 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ae89d00-a652-471e-bea0-1ce8c690e257-dns-svc\") pod \"3ae89d00-a652-471e-bea0-1ce8c690e257\" (UID: \"3ae89d00-a652-471e-bea0-1ce8c690e257\") " Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.761208 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ae89d00-a652-471e-bea0-1ce8c690e257-kube-api-access-2x5dt" (OuterVolumeSpecName: "kube-api-access-2x5dt") pod "3ae89d00-a652-471e-bea0-1ce8c690e257" (UID: "3ae89d00-a652-471e-bea0-1ce8c690e257"). InnerVolumeSpecName "kube-api-access-2x5dt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.770152 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ae89d00-a652-471e-bea0-1ce8c690e257-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "3ae89d00-a652-471e-bea0-1ce8c690e257" (UID: "3ae89d00-a652-471e-bea0-1ce8c690e257"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.770426 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.801205 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ae89d00-a652-471e-bea0-1ce8c690e257-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "3ae89d00-a652-471e-bea0-1ce8c690e257" (UID: "3ae89d00-a652-471e-bea0-1ce8c690e257"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.818495 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ae89d00-a652-471e-bea0-1ce8c690e257-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "3ae89d00-a652-471e-bea0-1ce8c690e257" (UID: "3ae89d00-a652-471e-bea0-1ce8c690e257"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.829234 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ae89d00-a652-471e-bea0-1ce8c690e257-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "3ae89d00-a652-471e-bea0-1ce8c690e257" (UID: "3ae89d00-a652-471e-bea0-1ce8c690e257"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.837214 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lfrp7\" (UniqueName: \"kubernetes.io/projected/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-kube-api-access-lfrp7\") pod \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\" (UID: \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\") " Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.837512 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-combined-ca-bundle\") pod \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\" (UID: \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\") " Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.837602 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-config-data\") pod \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\" (UID: \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\") " Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.837708 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-run-httpd\") pod \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\" (UID: \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\") " Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.837779 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-log-httpd\") pod \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\" (UID: \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\") " Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.837838 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-sg-core-conf-yaml\") pod \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\" (UID: \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\") " Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.837951 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-scripts\") pod \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\" (UID: \"eb56ccd5-8cd3-4da0-b11f-8600af8abf86\") " Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.838339 4967 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/3ae89d00-a652-471e-bea0-1ce8c690e257-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.838398 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2x5dt\" (UniqueName: \"kubernetes.io/projected/3ae89d00-a652-471e-bea0-1ce8c690e257-kube-api-access-2x5dt\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.838449 4967 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/3ae89d00-a652-471e-bea0-1ce8c690e257-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.838497 4967 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/3ae89d00-a652-471e-bea0-1ce8c690e257-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.838556 4967 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/3ae89d00-a652-471e-bea0-1ce8c690e257-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.838968 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3ae89d00-a652-471e-bea0-1ce8c690e257-config" (OuterVolumeSpecName: "config") pod "3ae89d00-a652-471e-bea0-1ce8c690e257" (UID: "3ae89d00-a652-471e-bea0-1ce8c690e257"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.839148 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "eb56ccd5-8cd3-4da0-b11f-8600af8abf86" (UID: "eb56ccd5-8cd3-4da0-b11f-8600af8abf86"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.840565 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-kube-api-access-lfrp7" (OuterVolumeSpecName: "kube-api-access-lfrp7") pod "eb56ccd5-8cd3-4da0-b11f-8600af8abf86" (UID: "eb56ccd5-8cd3-4da0-b11f-8600af8abf86"). InnerVolumeSpecName "kube-api-access-lfrp7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.840892 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "eb56ccd5-8cd3-4da0-b11f-8600af8abf86" (UID: "eb56ccd5-8cd3-4da0-b11f-8600af8abf86"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.848901 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-scripts" (OuterVolumeSpecName: "scripts") pod "eb56ccd5-8cd3-4da0-b11f-8600af8abf86" (UID: "eb56ccd5-8cd3-4da0-b11f-8600af8abf86"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.905706 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-jdrjp"] Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.940202 4967 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.940498 4967 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.940507 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.940515 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lfrp7\" (UniqueName: \"kubernetes.io/projected/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-kube-api-access-lfrp7\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.940525 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/3ae89d00-a652-471e-bea0-1ce8c690e257-config\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.946991 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "eb56ccd5-8cd3-4da0-b11f-8600af8abf86" (UID: "eb56ccd5-8cd3-4da0-b11f-8600af8abf86"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:18 crc kubenswrapper[4967]: I1011 04:08:18.970553 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eb56ccd5-8cd3-4da0-b11f-8600af8abf86" (UID: "eb56ccd5-8cd3-4da0-b11f-8600af8abf86"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.024270 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-config-data" (OuterVolumeSpecName: "config-data") pod "eb56ccd5-8cd3-4da0-b11f-8600af8abf86" (UID: "eb56ccd5-8cd3-4da0-b11f-8600af8abf86"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.042137 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.042169 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.042178 4967 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eb56ccd5-8cd3-4da0-b11f-8600af8abf86-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.083310 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-6554fc45b4-x7btb"] Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.119821 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-jdrjp" event={"ID":"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43","Type":"ContainerStarted","Data":"40529b23e0dd5d3fd6c984724438049d3aebc7de61e95e71d71b12869b2e206b"} Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.124931 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6554fc45b4-x7btb" event={"ID":"92205841-5c2e-4fc7-9328-8a164f5fb8cf","Type":"ContainerStarted","Data":"8689a5db47306a95c7d6f1322eaa82093ce6843fba7dbb3f3f9c41ee4440cbd9"} Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.126134 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-7c67bffd47-wgb47" event={"ID":"3ae89d00-a652-471e-bea0-1ce8c690e257","Type":"ContainerDied","Data":"435831d7c0933f1b223580056ad5882aa12216f78da38658b2772d88d12921a3"} Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.126166 4967 scope.go:117] "RemoveContainer" containerID="809c6e9eb61c4ab3be8c3e690ce8d39887b347035df7db22bfbfff76fdc1fe03" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.126279 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-7c67bffd47-wgb47" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.151283 4967 generic.go:334] "Generic (PLEG): container finished" podID="eb56ccd5-8cd3-4da0-b11f-8600af8abf86" containerID="ec8b91f055b4e4c15ad8ebb839b6d61ae69225f0fd74f6083d4bc346cf2ab43b" exitCode=0 Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.151376 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb56ccd5-8cd3-4da0-b11f-8600af8abf86","Type":"ContainerDied","Data":"ec8b91f055b4e4c15ad8ebb839b6d61ae69225f0fd74f6083d4bc346cf2ab43b"} Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.151407 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eb56ccd5-8cd3-4da0-b11f-8600af8abf86","Type":"ContainerDied","Data":"a8f828a9cd7f97607aa052ce81010e0423cf521eec734e13ca4ea8d5fb4fa18d"} Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.151491 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.195242 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b6c6c8584-4qwkb" event={"ID":"9c6c522c-bd96-42f6-b0ae-27d55d3121c4","Type":"ContainerStarted","Data":"fab128d00e2d02acf1adf0f691a5c9d7a4302c62e8039f94ba4a887ed5d74b69"} Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.197490 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6b6c6c8584-4qwkb" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.197526 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-6b6c6c8584-4qwkb" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.198059 4967 scope.go:117] "RemoveContainer" containerID="b73b0ac6958bcde79cb6d728ac0977bd7a8319a230cd00a4e2692a9e90ab183c" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.264245 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-7c67bffd47-wgb47"] Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.278546 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-7c67bffd47-wgb47"] Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.299925 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.305169 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.321375 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:08:19 crc kubenswrapper[4967]: E1011 04:08:19.321843 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb56ccd5-8cd3-4da0-b11f-8600af8abf86" containerName="ceilometer-notification-agent" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.321862 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb56ccd5-8cd3-4da0-b11f-8600af8abf86" containerName="ceilometer-notification-agent" Oct 11 04:08:19 crc kubenswrapper[4967]: E1011 04:08:19.321890 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3ae89d00-a652-471e-bea0-1ce8c690e257" containerName="init" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.321898 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="3ae89d00-a652-471e-bea0-1ce8c690e257" containerName="init" Oct 11 04:08:19 crc kubenswrapper[4967]: E1011 04:08:19.321910 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb56ccd5-8cd3-4da0-b11f-8600af8abf86" containerName="sg-core" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.321918 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb56ccd5-8cd3-4da0-b11f-8600af8abf86" containerName="sg-core" Oct 11 04:08:19 crc kubenswrapper[4967]: E1011 04:08:19.321937 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb56ccd5-8cd3-4da0-b11f-8600af8abf86" containerName="proxy-httpd" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.321944 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb56ccd5-8cd3-4da0-b11f-8600af8abf86" containerName="proxy-httpd" Oct 11 04:08:19 crc kubenswrapper[4967]: E1011 04:08:19.321961 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eb56ccd5-8cd3-4da0-b11f-8600af8abf86" containerName="ceilometer-central-agent" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.321970 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="eb56ccd5-8cd3-4da0-b11f-8600af8abf86" containerName="ceilometer-central-agent" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.322206 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb56ccd5-8cd3-4da0-b11f-8600af8abf86" containerName="ceilometer-notification-agent" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.322242 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb56ccd5-8cd3-4da0-b11f-8600af8abf86" containerName="proxy-httpd" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.322259 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="3ae89d00-a652-471e-bea0-1ce8c690e257" containerName="init" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.322275 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb56ccd5-8cd3-4da0-b11f-8600af8abf86" containerName="ceilometer-central-agent" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.322289 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="eb56ccd5-8cd3-4da0-b11f-8600af8abf86" containerName="sg-core" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.324439 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.326783 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-6b6c6c8584-4qwkb" podStartSLOduration=3.326769173 podStartE2EDuration="3.326769173s" podCreationTimestamp="2025-10-11 04:08:16 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:08:19.253480302 +0000 UTC m=+1027.216689235" watchObservedRunningTime="2025-10-11 04:08:19.326769173 +0000 UTC m=+1027.289978106" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.344394 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.345190 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.345255 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.367319 4967 scope.go:117] "RemoveContainer" containerID="41b3bcad4054fd47b10fc41d843fdb5fbed7ef7a180a89424017ebb74d7b6fa7" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.466016 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\") " pod="openstack/ceilometer-0" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.466090 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-config-data\") pod \"ceilometer-0\" (UID: \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\") " pod="openstack/ceilometer-0" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.466111 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-slqbp\" (UniqueName: \"kubernetes.io/projected/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-kube-api-access-slqbp\") pod \"ceilometer-0\" (UID: \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\") " pod="openstack/ceilometer-0" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.466132 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-scripts\") pod \"ceilometer-0\" (UID: \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\") " pod="openstack/ceilometer-0" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.466154 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\") " pod="openstack/ceilometer-0" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.466188 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-log-httpd\") pod \"ceilometer-0\" (UID: \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\") " pod="openstack/ceilometer-0" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.466217 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-run-httpd\") pod \"ceilometer-0\" (UID: \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\") " pod="openstack/ceilometer-0" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.524235 4967 scope.go:117] "RemoveContainer" containerID="ec8b91f055b4e4c15ad8ebb839b6d61ae69225f0fd74f6083d4bc346cf2ab43b" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.554449 4967 scope.go:117] "RemoveContainer" containerID="26117f8c594b515e35c7c2108fef08a9638b8113dcdf4be3945b880c53864c0d" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.568574 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-log-httpd\") pod \"ceilometer-0\" (UID: \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\") " pod="openstack/ceilometer-0" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.568649 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-run-httpd\") pod \"ceilometer-0\" (UID: \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\") " pod="openstack/ceilometer-0" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.568745 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\") " pod="openstack/ceilometer-0" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.568804 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-config-data\") pod \"ceilometer-0\" (UID: \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\") " pod="openstack/ceilometer-0" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.568835 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-slqbp\" (UniqueName: \"kubernetes.io/projected/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-kube-api-access-slqbp\") pod \"ceilometer-0\" (UID: \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\") " pod="openstack/ceilometer-0" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.568868 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-scripts\") pod \"ceilometer-0\" (UID: \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\") " pod="openstack/ceilometer-0" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.568902 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\") " pod="openstack/ceilometer-0" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.570884 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-log-httpd\") pod \"ceilometer-0\" (UID: \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\") " pod="openstack/ceilometer-0" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.571156 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-run-httpd\") pod \"ceilometer-0\" (UID: \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\") " pod="openstack/ceilometer-0" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.577854 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\") " pod="openstack/ceilometer-0" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.579619 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-config-data\") pod \"ceilometer-0\" (UID: \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\") " pod="openstack/ceilometer-0" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.581423 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-scripts\") pod \"ceilometer-0\" (UID: \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\") " pod="openstack/ceilometer-0" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.581578 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\") " pod="openstack/ceilometer-0" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.593901 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-slqbp\" (UniqueName: \"kubernetes.io/projected/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-kube-api-access-slqbp\") pod \"ceilometer-0\" (UID: \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\") " pod="openstack/ceilometer-0" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.664997 4967 scope.go:117] "RemoveContainer" containerID="b73b0ac6958bcde79cb6d728ac0977bd7a8319a230cd00a4e2692a9e90ab183c" Oct 11 04:08:19 crc kubenswrapper[4967]: E1011 04:08:19.665811 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b73b0ac6958bcde79cb6d728ac0977bd7a8319a230cd00a4e2692a9e90ab183c\": container with ID starting with b73b0ac6958bcde79cb6d728ac0977bd7a8319a230cd00a4e2692a9e90ab183c not found: ID does not exist" containerID="b73b0ac6958bcde79cb6d728ac0977bd7a8319a230cd00a4e2692a9e90ab183c" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.665911 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b73b0ac6958bcde79cb6d728ac0977bd7a8319a230cd00a4e2692a9e90ab183c"} err="failed to get container status \"b73b0ac6958bcde79cb6d728ac0977bd7a8319a230cd00a4e2692a9e90ab183c\": rpc error: code = NotFound desc = could not find container \"b73b0ac6958bcde79cb6d728ac0977bd7a8319a230cd00a4e2692a9e90ab183c\": container with ID starting with b73b0ac6958bcde79cb6d728ac0977bd7a8319a230cd00a4e2692a9e90ab183c not found: ID does not exist" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.665984 4967 scope.go:117] "RemoveContainer" containerID="41b3bcad4054fd47b10fc41d843fdb5fbed7ef7a180a89424017ebb74d7b6fa7" Oct 11 04:08:19 crc kubenswrapper[4967]: E1011 04:08:19.666368 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"41b3bcad4054fd47b10fc41d843fdb5fbed7ef7a180a89424017ebb74d7b6fa7\": container with ID starting with 41b3bcad4054fd47b10fc41d843fdb5fbed7ef7a180a89424017ebb74d7b6fa7 not found: ID does not exist" containerID="41b3bcad4054fd47b10fc41d843fdb5fbed7ef7a180a89424017ebb74d7b6fa7" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.666391 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"41b3bcad4054fd47b10fc41d843fdb5fbed7ef7a180a89424017ebb74d7b6fa7"} err="failed to get container status \"41b3bcad4054fd47b10fc41d843fdb5fbed7ef7a180a89424017ebb74d7b6fa7\": rpc error: code = NotFound desc = could not find container \"41b3bcad4054fd47b10fc41d843fdb5fbed7ef7a180a89424017ebb74d7b6fa7\": container with ID starting with 41b3bcad4054fd47b10fc41d843fdb5fbed7ef7a180a89424017ebb74d7b6fa7 not found: ID does not exist" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.666403 4967 scope.go:117] "RemoveContainer" containerID="ec8b91f055b4e4c15ad8ebb839b6d61ae69225f0fd74f6083d4bc346cf2ab43b" Oct 11 04:08:19 crc kubenswrapper[4967]: E1011 04:08:19.667736 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec8b91f055b4e4c15ad8ebb839b6d61ae69225f0fd74f6083d4bc346cf2ab43b\": container with ID starting with ec8b91f055b4e4c15ad8ebb839b6d61ae69225f0fd74f6083d4bc346cf2ab43b not found: ID does not exist" containerID="ec8b91f055b4e4c15ad8ebb839b6d61ae69225f0fd74f6083d4bc346cf2ab43b" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.667760 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec8b91f055b4e4c15ad8ebb839b6d61ae69225f0fd74f6083d4bc346cf2ab43b"} err="failed to get container status \"ec8b91f055b4e4c15ad8ebb839b6d61ae69225f0fd74f6083d4bc346cf2ab43b\": rpc error: code = NotFound desc = could not find container \"ec8b91f055b4e4c15ad8ebb839b6d61ae69225f0fd74f6083d4bc346cf2ab43b\": container with ID starting with ec8b91f055b4e4c15ad8ebb839b6d61ae69225f0fd74f6083d4bc346cf2ab43b not found: ID does not exist" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.667776 4967 scope.go:117] "RemoveContainer" containerID="26117f8c594b515e35c7c2108fef08a9638b8113dcdf4be3945b880c53864c0d" Oct 11 04:08:19 crc kubenswrapper[4967]: E1011 04:08:19.668082 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"26117f8c594b515e35c7c2108fef08a9638b8113dcdf4be3945b880c53864c0d\": container with ID starting with 26117f8c594b515e35c7c2108fef08a9638b8113dcdf4be3945b880c53864c0d not found: ID does not exist" containerID="26117f8c594b515e35c7c2108fef08a9638b8113dcdf4be3945b880c53864c0d" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.668107 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"26117f8c594b515e35c7c2108fef08a9638b8113dcdf4be3945b880c53864c0d"} err="failed to get container status \"26117f8c594b515e35c7c2108fef08a9638b8113dcdf4be3945b880c53864c0d\": rpc error: code = NotFound desc = could not find container \"26117f8c594b515e35c7c2108fef08a9638b8113dcdf4be3945b880c53864c0d\": container with ID starting with 26117f8c594b515e35c7c2108fef08a9638b8113dcdf4be3945b880c53864c0d not found: ID does not exist" Oct 11 04:08:19 crc kubenswrapper[4967]: I1011 04:08:19.809130 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:08:20 crc kubenswrapper[4967]: I1011 04:08:20.217243 4967 generic.go:334] "Generic (PLEG): container finished" podID="7a6e1897-5fd1-4c34-8f26-cd97ced1eb43" containerID="88cebd6157e8cdeb6552dc56736b5a329542c1553a121919af966eab72804d6d" exitCode=0 Oct 11 04:08:20 crc kubenswrapper[4967]: I1011 04:08:20.217542 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-jdrjp" event={"ID":"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43","Type":"ContainerDied","Data":"88cebd6157e8cdeb6552dc56736b5a329542c1553a121919af966eab72804d6d"} Oct 11 04:08:20 crc kubenswrapper[4967]: I1011 04:08:20.236264 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6554fc45b4-x7btb" event={"ID":"92205841-5c2e-4fc7-9328-8a164f5fb8cf","Type":"ContainerStarted","Data":"97c51df86213c4cd58189937091722374bf529014675fdfdb2cbbc6d2816f6dd"} Oct 11 04:08:20 crc kubenswrapper[4967]: I1011 04:08:20.411802 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:08:20 crc kubenswrapper[4967]: I1011 04:08:20.830645 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ae89d00-a652-471e-bea0-1ce8c690e257" path="/var/lib/kubelet/pods/3ae89d00-a652-471e-bea0-1ce8c690e257/volumes" Oct 11 04:08:20 crc kubenswrapper[4967]: I1011 04:08:20.831742 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eb56ccd5-8cd3-4da0-b11f-8600af8abf86" path="/var/lib/kubelet/pods/eb56ccd5-8cd3-4da0-b11f-8600af8abf86/volumes" Oct 11 04:08:20 crc kubenswrapper[4967]: I1011 04:08:20.872152 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/keystone-76b6798d84-nx4fq" Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.256626 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9d5e1373-95a7-4ec3-9316-c24cc9ee8833","Type":"ContainerStarted","Data":"1b1309c358fc7af4082c92f069e845f6d0876cf6f208f8c161141c49291ecd94"} Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.260041 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6554fc45b4-x7btb" event={"ID":"92205841-5c2e-4fc7-9328-8a164f5fb8cf","Type":"ContainerStarted","Data":"50f50943493197ea90b614621946c4164887b9cfb4c14fbe86b0b763dd2f42b6"} Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.260238 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-6554fc45b4-x7btb" Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.266458 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-jdrjp" event={"ID":"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43","Type":"ContainerStarted","Data":"21c9768d6219505f16dbf91b5cc918947d01146763a1e6c0d31c780ff517d5c9"} Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.266497 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-848cf88cfc-jdrjp" Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.274717 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5cdcf9b996-625cn" Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.275852 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/placement-5cdcf9b996-625cn" Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.293278 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-6554fc45b4-x7btb" podStartSLOduration=4.29325619 podStartE2EDuration="4.29325619s" podCreationTimestamp="2025-10-11 04:08:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:08:21.280956087 +0000 UTC m=+1029.244165020" watchObservedRunningTime="2025-10-11 04:08:21.29325619 +0000 UTC m=+1029.256465133" Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.313090 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-848cf88cfc-jdrjp" podStartSLOduration=4.313054906 podStartE2EDuration="4.313054906s" podCreationTimestamp="2025-10-11 04:08:17 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:08:21.305984813 +0000 UTC m=+1029.269193746" watchObservedRunningTime="2025-10-11 04:08:21.313054906 +0000 UTC m=+1029.276263849" Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.703627 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/neutron-5f96fd6969-6srgk"] Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.705499 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5f96fd6969-6srgk" Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.707443 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-public-svc" Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.710619 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-neutron-internal-svc" Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.712196 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5f96fd6969-6srgk"] Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.808110 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c21d8a72-9113-4b40-9cd2-fa7f56f749cd-combined-ca-bundle\") pod \"neutron-5f96fd6969-6srgk\" (UID: \"c21d8a72-9113-4b40-9cd2-fa7f56f749cd\") " pod="openstack/neutron-5f96fd6969-6srgk" Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.808171 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c21d8a72-9113-4b40-9cd2-fa7f56f749cd-public-tls-certs\") pod \"neutron-5f96fd6969-6srgk\" (UID: \"c21d8a72-9113-4b40-9cd2-fa7f56f749cd\") " pod="openstack/neutron-5f96fd6969-6srgk" Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.808221 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxwm5\" (UniqueName: \"kubernetes.io/projected/c21d8a72-9113-4b40-9cd2-fa7f56f749cd-kube-api-access-jxwm5\") pod \"neutron-5f96fd6969-6srgk\" (UID: \"c21d8a72-9113-4b40-9cd2-fa7f56f749cd\") " pod="openstack/neutron-5f96fd6969-6srgk" Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.808242 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c21d8a72-9113-4b40-9cd2-fa7f56f749cd-internal-tls-certs\") pod \"neutron-5f96fd6969-6srgk\" (UID: \"c21d8a72-9113-4b40-9cd2-fa7f56f749cd\") " pod="openstack/neutron-5f96fd6969-6srgk" Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.808277 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c21d8a72-9113-4b40-9cd2-fa7f56f749cd-config\") pod \"neutron-5f96fd6969-6srgk\" (UID: \"c21d8a72-9113-4b40-9cd2-fa7f56f749cd\") " pod="openstack/neutron-5f96fd6969-6srgk" Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.808314 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c21d8a72-9113-4b40-9cd2-fa7f56f749cd-ovndb-tls-certs\") pod \"neutron-5f96fd6969-6srgk\" (UID: \"c21d8a72-9113-4b40-9cd2-fa7f56f749cd\") " pod="openstack/neutron-5f96fd6969-6srgk" Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.808340 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c21d8a72-9113-4b40-9cd2-fa7f56f749cd-httpd-config\") pod \"neutron-5f96fd6969-6srgk\" (UID: \"c21d8a72-9113-4b40-9cd2-fa7f56f749cd\") " pod="openstack/neutron-5f96fd6969-6srgk" Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.910030 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c21d8a72-9113-4b40-9cd2-fa7f56f749cd-ovndb-tls-certs\") pod \"neutron-5f96fd6969-6srgk\" (UID: \"c21d8a72-9113-4b40-9cd2-fa7f56f749cd\") " pod="openstack/neutron-5f96fd6969-6srgk" Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.910600 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c21d8a72-9113-4b40-9cd2-fa7f56f749cd-httpd-config\") pod \"neutron-5f96fd6969-6srgk\" (UID: \"c21d8a72-9113-4b40-9cd2-fa7f56f749cd\") " pod="openstack/neutron-5f96fd6969-6srgk" Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.910869 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c21d8a72-9113-4b40-9cd2-fa7f56f749cd-combined-ca-bundle\") pod \"neutron-5f96fd6969-6srgk\" (UID: \"c21d8a72-9113-4b40-9cd2-fa7f56f749cd\") " pod="openstack/neutron-5f96fd6969-6srgk" Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.910974 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c21d8a72-9113-4b40-9cd2-fa7f56f749cd-public-tls-certs\") pod \"neutron-5f96fd6969-6srgk\" (UID: \"c21d8a72-9113-4b40-9cd2-fa7f56f749cd\") " pod="openstack/neutron-5f96fd6969-6srgk" Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.911120 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxwm5\" (UniqueName: \"kubernetes.io/projected/c21d8a72-9113-4b40-9cd2-fa7f56f749cd-kube-api-access-jxwm5\") pod \"neutron-5f96fd6969-6srgk\" (UID: \"c21d8a72-9113-4b40-9cd2-fa7f56f749cd\") " pod="openstack/neutron-5f96fd6969-6srgk" Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.911163 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c21d8a72-9113-4b40-9cd2-fa7f56f749cd-internal-tls-certs\") pod \"neutron-5f96fd6969-6srgk\" (UID: \"c21d8a72-9113-4b40-9cd2-fa7f56f749cd\") " pod="openstack/neutron-5f96fd6969-6srgk" Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.911247 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c21d8a72-9113-4b40-9cd2-fa7f56f749cd-config\") pod \"neutron-5f96fd6969-6srgk\" (UID: \"c21d8a72-9113-4b40-9cd2-fa7f56f749cd\") " pod="openstack/neutron-5f96fd6969-6srgk" Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.920743 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c21d8a72-9113-4b40-9cd2-fa7f56f749cd-internal-tls-certs\") pod \"neutron-5f96fd6969-6srgk\" (UID: \"c21d8a72-9113-4b40-9cd2-fa7f56f749cd\") " pod="openstack/neutron-5f96fd6969-6srgk" Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.925684 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c21d8a72-9113-4b40-9cd2-fa7f56f749cd-combined-ca-bundle\") pod \"neutron-5f96fd6969-6srgk\" (UID: \"c21d8a72-9113-4b40-9cd2-fa7f56f749cd\") " pod="openstack/neutron-5f96fd6969-6srgk" Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.926905 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/c21d8a72-9113-4b40-9cd2-fa7f56f749cd-config\") pod \"neutron-5f96fd6969-6srgk\" (UID: \"c21d8a72-9113-4b40-9cd2-fa7f56f749cd\") " pod="openstack/neutron-5f96fd6969-6srgk" Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.930645 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/c21d8a72-9113-4b40-9cd2-fa7f56f749cd-httpd-config\") pod \"neutron-5f96fd6969-6srgk\" (UID: \"c21d8a72-9113-4b40-9cd2-fa7f56f749cd\") " pod="openstack/neutron-5f96fd6969-6srgk" Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.930969 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c21d8a72-9113-4b40-9cd2-fa7f56f749cd-public-tls-certs\") pod \"neutron-5f96fd6969-6srgk\" (UID: \"c21d8a72-9113-4b40-9cd2-fa7f56f749cd\") " pod="openstack/neutron-5f96fd6969-6srgk" Oct 11 04:08:21 crc kubenswrapper[4967]: I1011 04:08:21.931996 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/c21d8a72-9113-4b40-9cd2-fa7f56f749cd-ovndb-tls-certs\") pod \"neutron-5f96fd6969-6srgk\" (UID: \"c21d8a72-9113-4b40-9cd2-fa7f56f749cd\") " pod="openstack/neutron-5f96fd6969-6srgk" Oct 11 04:08:22 crc kubenswrapper[4967]: I1011 04:08:22.005102 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxwm5\" (UniqueName: \"kubernetes.io/projected/c21d8a72-9113-4b40-9cd2-fa7f56f749cd-kube-api-access-jxwm5\") pod \"neutron-5f96fd6969-6srgk\" (UID: \"c21d8a72-9113-4b40-9cd2-fa7f56f749cd\") " pod="openstack/neutron-5f96fd6969-6srgk" Oct 11 04:08:22 crc kubenswrapper[4967]: I1011 04:08:22.030411 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-5f96fd6969-6srgk" Oct 11 04:08:22 crc kubenswrapper[4967]: I1011 04:08:22.896625 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/barbican-api-84674df998-htlks"] Oct 11 04:08:22 crc kubenswrapper[4967]: I1011 04:08:22.898167 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-84674df998-htlks" Oct 11 04:08:22 crc kubenswrapper[4967]: I1011 04:08:22.903575 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-internal-svc" Oct 11 04:08:22 crc kubenswrapper[4967]: I1011 04:08:22.903780 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-barbican-public-svc" Oct 11 04:08:22 crc kubenswrapper[4967]: I1011 04:08:22.912334 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-84674df998-htlks"] Oct 11 04:08:23 crc kubenswrapper[4967]: I1011 04:08:23.044523 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8d6kq\" (UniqueName: \"kubernetes.io/projected/c72a2b43-59ac-450c-87d3-f16c9ee4afa0-kube-api-access-8d6kq\") pod \"barbican-api-84674df998-htlks\" (UID: \"c72a2b43-59ac-450c-87d3-f16c9ee4afa0\") " pod="openstack/barbican-api-84674df998-htlks" Oct 11 04:08:23 crc kubenswrapper[4967]: I1011 04:08:23.044562 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c72a2b43-59ac-450c-87d3-f16c9ee4afa0-config-data-custom\") pod \"barbican-api-84674df998-htlks\" (UID: \"c72a2b43-59ac-450c-87d3-f16c9ee4afa0\") " pod="openstack/barbican-api-84674df998-htlks" Oct 11 04:08:23 crc kubenswrapper[4967]: I1011 04:08:23.044603 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c72a2b43-59ac-450c-87d3-f16c9ee4afa0-config-data\") pod \"barbican-api-84674df998-htlks\" (UID: \"c72a2b43-59ac-450c-87d3-f16c9ee4afa0\") " pod="openstack/barbican-api-84674df998-htlks" Oct 11 04:08:23 crc kubenswrapper[4967]: I1011 04:08:23.044629 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c72a2b43-59ac-450c-87d3-f16c9ee4afa0-public-tls-certs\") pod \"barbican-api-84674df998-htlks\" (UID: \"c72a2b43-59ac-450c-87d3-f16c9ee4afa0\") " pod="openstack/barbican-api-84674df998-htlks" Oct 11 04:08:23 crc kubenswrapper[4967]: I1011 04:08:23.044683 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c72a2b43-59ac-450c-87d3-f16c9ee4afa0-logs\") pod \"barbican-api-84674df998-htlks\" (UID: \"c72a2b43-59ac-450c-87d3-f16c9ee4afa0\") " pod="openstack/barbican-api-84674df998-htlks" Oct 11 04:08:23 crc kubenswrapper[4967]: I1011 04:08:23.045258 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c72a2b43-59ac-450c-87d3-f16c9ee4afa0-internal-tls-certs\") pod \"barbican-api-84674df998-htlks\" (UID: \"c72a2b43-59ac-450c-87d3-f16c9ee4afa0\") " pod="openstack/barbican-api-84674df998-htlks" Oct 11 04:08:23 crc kubenswrapper[4967]: I1011 04:08:23.045321 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c72a2b43-59ac-450c-87d3-f16c9ee4afa0-combined-ca-bundle\") pod \"barbican-api-84674df998-htlks\" (UID: \"c72a2b43-59ac-450c-87d3-f16c9ee4afa0\") " pod="openstack/barbican-api-84674df998-htlks" Oct 11 04:08:23 crc kubenswrapper[4967]: I1011 04:08:23.146496 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8d6kq\" (UniqueName: \"kubernetes.io/projected/c72a2b43-59ac-450c-87d3-f16c9ee4afa0-kube-api-access-8d6kq\") pod \"barbican-api-84674df998-htlks\" (UID: \"c72a2b43-59ac-450c-87d3-f16c9ee4afa0\") " pod="openstack/barbican-api-84674df998-htlks" Oct 11 04:08:23 crc kubenswrapper[4967]: I1011 04:08:23.146548 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c72a2b43-59ac-450c-87d3-f16c9ee4afa0-config-data-custom\") pod \"barbican-api-84674df998-htlks\" (UID: \"c72a2b43-59ac-450c-87d3-f16c9ee4afa0\") " pod="openstack/barbican-api-84674df998-htlks" Oct 11 04:08:23 crc kubenswrapper[4967]: I1011 04:08:23.146596 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c72a2b43-59ac-450c-87d3-f16c9ee4afa0-config-data\") pod \"barbican-api-84674df998-htlks\" (UID: \"c72a2b43-59ac-450c-87d3-f16c9ee4afa0\") " pod="openstack/barbican-api-84674df998-htlks" Oct 11 04:08:23 crc kubenswrapper[4967]: I1011 04:08:23.146625 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c72a2b43-59ac-450c-87d3-f16c9ee4afa0-public-tls-certs\") pod \"barbican-api-84674df998-htlks\" (UID: \"c72a2b43-59ac-450c-87d3-f16c9ee4afa0\") " pod="openstack/barbican-api-84674df998-htlks" Oct 11 04:08:23 crc kubenswrapper[4967]: I1011 04:08:23.146695 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c72a2b43-59ac-450c-87d3-f16c9ee4afa0-logs\") pod \"barbican-api-84674df998-htlks\" (UID: \"c72a2b43-59ac-450c-87d3-f16c9ee4afa0\") " pod="openstack/barbican-api-84674df998-htlks" Oct 11 04:08:23 crc kubenswrapper[4967]: I1011 04:08:23.146713 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c72a2b43-59ac-450c-87d3-f16c9ee4afa0-internal-tls-certs\") pod \"barbican-api-84674df998-htlks\" (UID: \"c72a2b43-59ac-450c-87d3-f16c9ee4afa0\") " pod="openstack/barbican-api-84674df998-htlks" Oct 11 04:08:23 crc kubenswrapper[4967]: I1011 04:08:23.150446 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c72a2b43-59ac-450c-87d3-f16c9ee4afa0-combined-ca-bundle\") pod \"barbican-api-84674df998-htlks\" (UID: \"c72a2b43-59ac-450c-87d3-f16c9ee4afa0\") " pod="openstack/barbican-api-84674df998-htlks" Oct 11 04:08:23 crc kubenswrapper[4967]: I1011 04:08:23.148498 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c72a2b43-59ac-450c-87d3-f16c9ee4afa0-logs\") pod \"barbican-api-84674df998-htlks\" (UID: \"c72a2b43-59ac-450c-87d3-f16c9ee4afa0\") " pod="openstack/barbican-api-84674df998-htlks" Oct 11 04:08:23 crc kubenswrapper[4967]: I1011 04:08:23.154487 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c72a2b43-59ac-450c-87d3-f16c9ee4afa0-config-data\") pod \"barbican-api-84674df998-htlks\" (UID: \"c72a2b43-59ac-450c-87d3-f16c9ee4afa0\") " pod="openstack/barbican-api-84674df998-htlks" Oct 11 04:08:23 crc kubenswrapper[4967]: I1011 04:08:23.158500 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c72a2b43-59ac-450c-87d3-f16c9ee4afa0-internal-tls-certs\") pod \"barbican-api-84674df998-htlks\" (UID: \"c72a2b43-59ac-450c-87d3-f16c9ee4afa0\") " pod="openstack/barbican-api-84674df998-htlks" Oct 11 04:08:23 crc kubenswrapper[4967]: I1011 04:08:23.162590 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c72a2b43-59ac-450c-87d3-f16c9ee4afa0-public-tls-certs\") pod \"barbican-api-84674df998-htlks\" (UID: \"c72a2b43-59ac-450c-87d3-f16c9ee4afa0\") " pod="openstack/barbican-api-84674df998-htlks" Oct 11 04:08:23 crc kubenswrapper[4967]: I1011 04:08:23.166551 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8d6kq\" (UniqueName: \"kubernetes.io/projected/c72a2b43-59ac-450c-87d3-f16c9ee4afa0-kube-api-access-8d6kq\") pod \"barbican-api-84674df998-htlks\" (UID: \"c72a2b43-59ac-450c-87d3-f16c9ee4afa0\") " pod="openstack/barbican-api-84674df998-htlks" Oct 11 04:08:23 crc kubenswrapper[4967]: I1011 04:08:23.167689 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c72a2b43-59ac-450c-87d3-f16c9ee4afa0-combined-ca-bundle\") pod \"barbican-api-84674df998-htlks\" (UID: \"c72a2b43-59ac-450c-87d3-f16c9ee4afa0\") " pod="openstack/barbican-api-84674df998-htlks" Oct 11 04:08:23 crc kubenswrapper[4967]: I1011 04:08:23.167791 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c72a2b43-59ac-450c-87d3-f16c9ee4afa0-config-data-custom\") pod \"barbican-api-84674df998-htlks\" (UID: \"c72a2b43-59ac-450c-87d3-f16c9ee4afa0\") " pod="openstack/barbican-api-84674df998-htlks" Oct 11 04:08:23 crc kubenswrapper[4967]: I1011 04:08:23.445224 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-84674df998-htlks" Oct 11 04:08:23 crc kubenswrapper[4967]: I1011 04:08:23.483854 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/neutron-5f96fd6969-6srgk"] Oct 11 04:08:24 crc kubenswrapper[4967]: I1011 04:08:24.015594 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/barbican-api-84674df998-htlks"] Oct 11 04:08:24 crc kubenswrapper[4967]: I1011 04:08:24.077334 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6b6c6c8584-4qwkb" Oct 11 04:08:24 crc kubenswrapper[4967]: I1011 04:08:24.295620 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9d5e1373-95a7-4ec3-9316-c24cc9ee8833","Type":"ContainerStarted","Data":"dae619182846b7310ad63ec9ee260e7711654bf0e8b68976b12a1d8fdcad126c"} Oct 11 04:08:24 crc kubenswrapper[4967]: I1011 04:08:24.300059 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-84674df998-htlks" event={"ID":"c72a2b43-59ac-450c-87d3-f16c9ee4afa0","Type":"ContainerStarted","Data":"a5f71b769398c49a861453e9423dccc27756284707788a398334f9c527ce55f0"} Oct 11 04:08:24 crc kubenswrapper[4967]: I1011 04:08:24.313085 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5f96fd6969-6srgk" event={"ID":"c21d8a72-9113-4b40-9cd2-fa7f56f749cd","Type":"ContainerStarted","Data":"092ce2525eca6fb3b4b45c6dc080bc9b69c8561373ca24d670db1ebbd2692a77"} Oct 11 04:08:24 crc kubenswrapper[4967]: I1011 04:08:24.313129 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5f96fd6969-6srgk" event={"ID":"c21d8a72-9113-4b40-9cd2-fa7f56f749cd","Type":"ContainerStarted","Data":"9ff8a0f72d071a0cd7204e2ba24366431f041dcf7b12c4642ff0678467cfe405"} Oct 11 04:08:24 crc kubenswrapper[4967]: I1011 04:08:24.321776 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6c87f65cff-p6j2x" event={"ID":"fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35","Type":"ContainerStarted","Data":"1869821859b6b92fabef2e48e09e55d7f2ee7f241728c4f93d3466f3c3dccc63"} Oct 11 04:08:24 crc kubenswrapper[4967]: I1011 04:08:24.321844 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-worker-6c87f65cff-p6j2x" event={"ID":"fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35","Type":"ContainerStarted","Data":"d498d3ef3041167e686a7ebb477d43245b8046c4fca4eea6c65f16a795a65adf"} Oct 11 04:08:24 crc kubenswrapper[4967]: I1011 04:08:24.329495 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7769b9cccd-zkh5v" event={"ID":"38e5cf89-c793-4207-b06e-d117da3dd152","Type":"ContainerStarted","Data":"d996db6fe29b39f5d9dd66b88723ac60dc3ef8b0f50297fdb4b095d35e5b7d52"} Oct 11 04:08:24 crc kubenswrapper[4967]: I1011 04:08:24.329535 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-keystone-listener-7769b9cccd-zkh5v" event={"ID":"38e5cf89-c793-4207-b06e-d117da3dd152","Type":"ContainerStarted","Data":"c0f21c1dfe37157c5063164530a529abaef858b4c57b9aaa149b3884201f949c"} Oct 11 04:08:24 crc kubenswrapper[4967]: I1011 04:08:24.353987 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-worker-6c87f65cff-p6j2x" podStartSLOduration=3.007324195 podStartE2EDuration="8.35396202s" podCreationTimestamp="2025-10-11 04:08:16 +0000 UTC" firstStartedPulling="2025-10-11 04:08:17.434708801 +0000 UTC m=+1025.397917744" lastFinishedPulling="2025-10-11 04:08:22.781346616 +0000 UTC m=+1030.744555569" observedRunningTime="2025-10-11 04:08:24.349132819 +0000 UTC m=+1032.312341772" watchObservedRunningTime="2025-10-11 04:08:24.35396202 +0000 UTC m=+1032.317170973" Oct 11 04:08:24 crc kubenswrapper[4967]: I1011 04:08:24.372190 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-keystone-listener-7769b9cccd-zkh5v" podStartSLOduration=3.173433756 podStartE2EDuration="8.37217262s" podCreationTimestamp="2025-10-11 04:08:16 +0000 UTC" firstStartedPulling="2025-10-11 04:08:17.564624987 +0000 UTC m=+1025.527833920" lastFinishedPulling="2025-10-11 04:08:22.763363851 +0000 UTC m=+1030.726572784" observedRunningTime="2025-10-11 04:08:24.365388893 +0000 UTC m=+1032.328597826" watchObservedRunningTime="2025-10-11 04:08:24.37217262 +0000 UTC m=+1032.335381553" Oct 11 04:08:25 crc kubenswrapper[4967]: I1011 04:08:25.340627 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-84674df998-htlks" event={"ID":"c72a2b43-59ac-450c-87d3-f16c9ee4afa0","Type":"ContainerStarted","Data":"cb67727f88b7b6491be0eab9d5090ad2ca94b5f24d3f91b6d12116e2beabdbba"} Oct 11 04:08:25 crc kubenswrapper[4967]: I1011 04:08:25.340956 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-84674df998-htlks" event={"ID":"c72a2b43-59ac-450c-87d3-f16c9ee4afa0","Type":"ContainerStarted","Data":"a120c2d89fc2730a35caa651647c6c668521f9b61f28643da5056df752e30a4d"} Oct 11 04:08:25 crc kubenswrapper[4967]: I1011 04:08:25.341347 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-84674df998-htlks" Oct 11 04:08:25 crc kubenswrapper[4967]: I1011 04:08:25.342535 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-5f96fd6969-6srgk" event={"ID":"c21d8a72-9113-4b40-9cd2-fa7f56f749cd","Type":"ContainerStarted","Data":"2cd6533af2670dd9b2be343c4888c11bc8489af7638a9152b309ff9d6da8427a"} Oct 11 04:08:25 crc kubenswrapper[4967]: I1011 04:08:25.342680 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/neutron-5f96fd6969-6srgk" Oct 11 04:08:25 crc kubenswrapper[4967]: I1011 04:08:25.344332 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9d5e1373-95a7-4ec3-9316-c24cc9ee8833","Type":"ContainerStarted","Data":"7236195b1efa93e02f72597cbd63168d14fddc957d15a80644b2c0d778880cb6"} Oct 11 04:08:25 crc kubenswrapper[4967]: I1011 04:08:25.344640 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9d5e1373-95a7-4ec3-9316-c24cc9ee8833","Type":"ContainerStarted","Data":"c4d97530d947be64f12582efe3328e97cc2041a7c4b16e0fd87b9768a2dae2f0"} Oct 11 04:08:25 crc kubenswrapper[4967]: I1011 04:08:25.374142 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/barbican-api-84674df998-htlks" podStartSLOduration=3.374111015 podStartE2EDuration="3.374111015s" podCreationTimestamp="2025-10-11 04:08:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:08:25.361494354 +0000 UTC m=+1033.324703287" watchObservedRunningTime="2025-10-11 04:08:25.374111015 +0000 UTC m=+1033.337319958" Oct 11 04:08:25 crc kubenswrapper[4967]: I1011 04:08:25.824777 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/neutron-5f96fd6969-6srgk" podStartSLOduration=4.824759447 podStartE2EDuration="4.824759447s" podCreationTimestamp="2025-10-11 04:08:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:08:25.402760756 +0000 UTC m=+1033.365969689" watchObservedRunningTime="2025-10-11 04:08:25.824759447 +0000 UTC m=+1033.787968370" Oct 11 04:08:25 crc kubenswrapper[4967]: I1011 04:08:25.833324 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/openstackclient"] Oct 11 04:08:25 crc kubenswrapper[4967]: I1011 04:08:25.835187 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 11 04:08:25 crc kubenswrapper[4967]: I1011 04:08:25.837252 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstackclient-openstackclient-dockercfg-ml867" Oct 11 04:08:25 crc kubenswrapper[4967]: I1011 04:08:25.838548 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"openstack-config" Oct 11 04:08:25 crc kubenswrapper[4967]: I1011 04:08:25.838688 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"openstack-config-secret" Oct 11 04:08:25 crc kubenswrapper[4967]: I1011 04:08:25.845488 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 11 04:08:26 crc kubenswrapper[4967]: I1011 04:08:26.016509 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-6b6c6c8584-4qwkb" Oct 11 04:08:26 crc kubenswrapper[4967]: I1011 04:08:26.032325 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/23c5100e-07d0-4897-98de-f4e91329085e-openstack-config\") pod \"openstackclient\" (UID: \"23c5100e-07d0-4897-98de-f4e91329085e\") " pod="openstack/openstackclient" Oct 11 04:08:26 crc kubenswrapper[4967]: I1011 04:08:26.032384 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b2xn4\" (UniqueName: \"kubernetes.io/projected/23c5100e-07d0-4897-98de-f4e91329085e-kube-api-access-b2xn4\") pod \"openstackclient\" (UID: \"23c5100e-07d0-4897-98de-f4e91329085e\") " pod="openstack/openstackclient" Oct 11 04:08:26 crc kubenswrapper[4967]: I1011 04:08:26.032415 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23c5100e-07d0-4897-98de-f4e91329085e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"23c5100e-07d0-4897-98de-f4e91329085e\") " pod="openstack/openstackclient" Oct 11 04:08:26 crc kubenswrapper[4967]: I1011 04:08:26.032465 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/23c5100e-07d0-4897-98de-f4e91329085e-openstack-config-secret\") pod \"openstackclient\" (UID: \"23c5100e-07d0-4897-98de-f4e91329085e\") " pod="openstack/openstackclient" Oct 11 04:08:26 crc kubenswrapper[4967]: I1011 04:08:26.134539 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/23c5100e-07d0-4897-98de-f4e91329085e-openstack-config-secret\") pod \"openstackclient\" (UID: \"23c5100e-07d0-4897-98de-f4e91329085e\") " pod="openstack/openstackclient" Oct 11 04:08:26 crc kubenswrapper[4967]: I1011 04:08:26.134715 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/23c5100e-07d0-4897-98de-f4e91329085e-openstack-config\") pod \"openstackclient\" (UID: \"23c5100e-07d0-4897-98de-f4e91329085e\") " pod="openstack/openstackclient" Oct 11 04:08:26 crc kubenswrapper[4967]: I1011 04:08:26.134786 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b2xn4\" (UniqueName: \"kubernetes.io/projected/23c5100e-07d0-4897-98de-f4e91329085e-kube-api-access-b2xn4\") pod \"openstackclient\" (UID: \"23c5100e-07d0-4897-98de-f4e91329085e\") " pod="openstack/openstackclient" Oct 11 04:08:26 crc kubenswrapper[4967]: I1011 04:08:26.134829 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23c5100e-07d0-4897-98de-f4e91329085e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"23c5100e-07d0-4897-98de-f4e91329085e\") " pod="openstack/openstackclient" Oct 11 04:08:26 crc kubenswrapper[4967]: I1011 04:08:26.137891 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config\" (UniqueName: \"kubernetes.io/configmap/23c5100e-07d0-4897-98de-f4e91329085e-openstack-config\") pod \"openstackclient\" (UID: \"23c5100e-07d0-4897-98de-f4e91329085e\") " pod="openstack/openstackclient" Oct 11 04:08:26 crc kubenswrapper[4967]: I1011 04:08:26.142741 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openstack-config-secret\" (UniqueName: \"kubernetes.io/secret/23c5100e-07d0-4897-98de-f4e91329085e-openstack-config-secret\") pod \"openstackclient\" (UID: \"23c5100e-07d0-4897-98de-f4e91329085e\") " pod="openstack/openstackclient" Oct 11 04:08:26 crc kubenswrapper[4967]: I1011 04:08:26.143936 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/23c5100e-07d0-4897-98de-f4e91329085e-combined-ca-bundle\") pod \"openstackclient\" (UID: \"23c5100e-07d0-4897-98de-f4e91329085e\") " pod="openstack/openstackclient" Oct 11 04:08:26 crc kubenswrapper[4967]: I1011 04:08:26.160335 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b2xn4\" (UniqueName: \"kubernetes.io/projected/23c5100e-07d0-4897-98de-f4e91329085e-kube-api-access-b2xn4\") pod \"openstackclient\" (UID: \"23c5100e-07d0-4897-98de-f4e91329085e\") " pod="openstack/openstackclient" Oct 11 04:08:26 crc kubenswrapper[4967]: I1011 04:08:26.354691 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/barbican-api-84674df998-htlks" Oct 11 04:08:26 crc kubenswrapper[4967]: I1011 04:08:26.454799 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/openstackclient" Oct 11 04:08:27 crc kubenswrapper[4967]: I1011 04:08:27.103911 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/openstackclient"] Oct 11 04:08:27 crc kubenswrapper[4967]: W1011 04:08:27.304164 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod23c5100e_07d0_4897_98de_f4e91329085e.slice/crio-ce38f48cfebe31f4ab6d614b43ce0d2984b6ec8763d12f57b478588b2b060223 WatchSource:0}: Error finding container ce38f48cfebe31f4ab6d614b43ce0d2984b6ec8763d12f57b478588b2b060223: Status 404 returned error can't find the container with id ce38f48cfebe31f4ab6d614b43ce0d2984b6ec8763d12f57b478588b2b060223 Oct 11 04:08:27 crc kubenswrapper[4967]: I1011 04:08:27.363172 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"23c5100e-07d0-4897-98de-f4e91329085e","Type":"ContainerStarted","Data":"ce38f48cfebe31f4ab6d614b43ce0d2984b6ec8763d12f57b478588b2b060223"} Oct 11 04:08:28 crc kubenswrapper[4967]: I1011 04:08:28.145643 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-848cf88cfc-jdrjp" Oct 11 04:08:28 crc kubenswrapper[4967]: I1011 04:08:28.212618 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-k8f6p"] Oct 11 04:08:28 crc kubenswrapper[4967]: I1011 04:08:28.212916 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-56df8fb6b7-k8f6p" podUID="dfc995c0-dbf3-45a0-b16e-3d86acabe454" containerName="dnsmasq-dns" containerID="cri-o://1853439d5d6a6d705b66fdd687606d349872991825f0476ad70d3f6f597aded2" gracePeriod=10 Oct 11 04:08:28 crc kubenswrapper[4967]: I1011 04:08:28.393360 4967 generic.go:334] "Generic (PLEG): container finished" podID="dfc995c0-dbf3-45a0-b16e-3d86acabe454" containerID="1853439d5d6a6d705b66fdd687606d349872991825f0476ad70d3f6f597aded2" exitCode=0 Oct 11 04:08:28 crc kubenswrapper[4967]: I1011 04:08:28.393409 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-k8f6p" event={"ID":"dfc995c0-dbf3-45a0-b16e-3d86acabe454","Type":"ContainerDied","Data":"1853439d5d6a6d705b66fdd687606d349872991825f0476ad70d3f6f597aded2"} Oct 11 04:08:28 crc kubenswrapper[4967]: I1011 04:08:28.396000 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9d5e1373-95a7-4ec3-9316-c24cc9ee8833","Type":"ContainerStarted","Data":"f64b25630366a663ac416ada48352e991a15a0213371220d837e72ba513411e1"} Oct 11 04:08:28 crc kubenswrapper[4967]: I1011 04:08:28.396200 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 11 04:08:28 crc kubenswrapper[4967]: I1011 04:08:28.447495 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.275904121 podStartE2EDuration="9.447474271s" podCreationTimestamp="2025-10-11 04:08:19 +0000 UTC" firstStartedPulling="2025-10-11 04:08:20.434483067 +0000 UTC m=+1028.397692000" lastFinishedPulling="2025-10-11 04:08:27.606053217 +0000 UTC m=+1035.569262150" observedRunningTime="2025-10-11 04:08:28.433941191 +0000 UTC m=+1036.397150124" watchObservedRunningTime="2025-10-11 04:08:28.447474271 +0000 UTC m=+1036.410683204" Oct 11 04:08:28 crc kubenswrapper[4967]: I1011 04:08:28.962481 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-k8f6p" Oct 11 04:08:29 crc kubenswrapper[4967]: I1011 04:08:29.128806 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dfc995c0-dbf3-45a0-b16e-3d86acabe454-ovsdbserver-nb\") pod \"dfc995c0-dbf3-45a0-b16e-3d86acabe454\" (UID: \"dfc995c0-dbf3-45a0-b16e-3d86acabe454\") " Oct 11 04:08:29 crc kubenswrapper[4967]: I1011 04:08:29.128895 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dfc995c0-dbf3-45a0-b16e-3d86acabe454-ovsdbserver-sb\") pod \"dfc995c0-dbf3-45a0-b16e-3d86acabe454\" (UID: \"dfc995c0-dbf3-45a0-b16e-3d86acabe454\") " Oct 11 04:08:29 crc kubenswrapper[4967]: I1011 04:08:29.128979 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dfc995c0-dbf3-45a0-b16e-3d86acabe454-dns-swift-storage-0\") pod \"dfc995c0-dbf3-45a0-b16e-3d86acabe454\" (UID: \"dfc995c0-dbf3-45a0-b16e-3d86acabe454\") " Oct 11 04:08:29 crc kubenswrapper[4967]: I1011 04:08:29.129006 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfc995c0-dbf3-45a0-b16e-3d86acabe454-config\") pod \"dfc995c0-dbf3-45a0-b16e-3d86acabe454\" (UID: \"dfc995c0-dbf3-45a0-b16e-3d86acabe454\") " Oct 11 04:08:29 crc kubenswrapper[4967]: I1011 04:08:29.129082 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kkrxv\" (UniqueName: \"kubernetes.io/projected/dfc995c0-dbf3-45a0-b16e-3d86acabe454-kube-api-access-kkrxv\") pod \"dfc995c0-dbf3-45a0-b16e-3d86acabe454\" (UID: \"dfc995c0-dbf3-45a0-b16e-3d86acabe454\") " Oct 11 04:08:29 crc kubenswrapper[4967]: I1011 04:08:29.129133 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dfc995c0-dbf3-45a0-b16e-3d86acabe454-dns-svc\") pod \"dfc995c0-dbf3-45a0-b16e-3d86acabe454\" (UID: \"dfc995c0-dbf3-45a0-b16e-3d86acabe454\") " Oct 11 04:08:29 crc kubenswrapper[4967]: I1011 04:08:29.165580 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dfc995c0-dbf3-45a0-b16e-3d86acabe454-kube-api-access-kkrxv" (OuterVolumeSpecName: "kube-api-access-kkrxv") pod "dfc995c0-dbf3-45a0-b16e-3d86acabe454" (UID: "dfc995c0-dbf3-45a0-b16e-3d86acabe454"). InnerVolumeSpecName "kube-api-access-kkrxv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:08:29 crc kubenswrapper[4967]: I1011 04:08:29.222026 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfc995c0-dbf3-45a0-b16e-3d86acabe454-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "dfc995c0-dbf3-45a0-b16e-3d86acabe454" (UID: "dfc995c0-dbf3-45a0-b16e-3d86acabe454"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:08:29 crc kubenswrapper[4967]: I1011 04:08:29.232659 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfc995c0-dbf3-45a0-b16e-3d86acabe454-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "dfc995c0-dbf3-45a0-b16e-3d86acabe454" (UID: "dfc995c0-dbf3-45a0-b16e-3d86acabe454"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:08:29 crc kubenswrapper[4967]: I1011 04:08:29.232320 4967 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/dfc995c0-dbf3-45a0-b16e-3d86acabe454-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:29 crc kubenswrapper[4967]: I1011 04:08:29.232968 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kkrxv\" (UniqueName: \"kubernetes.io/projected/dfc995c0-dbf3-45a0-b16e-3d86acabe454-kube-api-access-kkrxv\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:29 crc kubenswrapper[4967]: I1011 04:08:29.259799 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfc995c0-dbf3-45a0-b16e-3d86acabe454-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "dfc995c0-dbf3-45a0-b16e-3d86acabe454" (UID: "dfc995c0-dbf3-45a0-b16e-3d86acabe454"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:08:29 crc kubenswrapper[4967]: I1011 04:08:29.285218 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfc995c0-dbf3-45a0-b16e-3d86acabe454-config" (OuterVolumeSpecName: "config") pod "dfc995c0-dbf3-45a0-b16e-3d86acabe454" (UID: "dfc995c0-dbf3-45a0-b16e-3d86acabe454"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:08:29 crc kubenswrapper[4967]: I1011 04:08:29.300168 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/dfc995c0-dbf3-45a0-b16e-3d86acabe454-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "dfc995c0-dbf3-45a0-b16e-3d86acabe454" (UID: "dfc995c0-dbf3-45a0-b16e-3d86acabe454"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:08:29 crc kubenswrapper[4967]: I1011 04:08:29.334224 4967 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/dfc995c0-dbf3-45a0-b16e-3d86acabe454-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:29 crc kubenswrapper[4967]: I1011 04:08:29.334256 4967 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/dfc995c0-dbf3-45a0-b16e-3d86acabe454-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:29 crc kubenswrapper[4967]: I1011 04:08:29.334268 4967 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/dfc995c0-dbf3-45a0-b16e-3d86acabe454-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:29 crc kubenswrapper[4967]: I1011 04:08:29.334279 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/dfc995c0-dbf3-45a0-b16e-3d86acabe454-config\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:29 crc kubenswrapper[4967]: I1011 04:08:29.406294 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-56df8fb6b7-k8f6p" event={"ID":"dfc995c0-dbf3-45a0-b16e-3d86acabe454","Type":"ContainerDied","Data":"b7edf5c52368d51a6e2f60b140c3b33189073dc358e1d1a4bc9494d28329c453"} Oct 11 04:08:29 crc kubenswrapper[4967]: I1011 04:08:29.406353 4967 scope.go:117] "RemoveContainer" containerID="1853439d5d6a6d705b66fdd687606d349872991825f0476ad70d3f6f597aded2" Oct 11 04:08:29 crc kubenswrapper[4967]: I1011 04:08:29.406466 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-56df8fb6b7-k8f6p" Oct 11 04:08:29 crc kubenswrapper[4967]: I1011 04:08:29.438211 4967 scope.go:117] "RemoveContainer" containerID="7fb30d62397b4f0465dd75c26bfec185771fabfeee3c3279a171d97a3ffa5e8c" Oct 11 04:08:29 crc kubenswrapper[4967]: I1011 04:08:29.451146 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-k8f6p"] Oct 11 04:08:29 crc kubenswrapper[4967]: I1011 04:08:29.487810 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-56df8fb6b7-k8f6p"] Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.658253 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/swift-proxy-6486584599-dfqxj"] Oct 11 04:08:30 crc kubenswrapper[4967]: E1011 04:08:30.658770 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfc995c0-dbf3-45a0-b16e-3d86acabe454" containerName="dnsmasq-dns" Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.658781 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfc995c0-dbf3-45a0-b16e-3d86acabe454" containerName="dnsmasq-dns" Oct 11 04:08:30 crc kubenswrapper[4967]: E1011 04:08:30.658798 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dfc995c0-dbf3-45a0-b16e-3d86acabe454" containerName="init" Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.658803 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="dfc995c0-dbf3-45a0-b16e-3d86acabe454" containerName="init" Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.658970 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="dfc995c0-dbf3-45a0-b16e-3d86acabe454" containerName="dnsmasq-dns" Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.659872 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6486584599-dfqxj" Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.667063 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-internal-svc" Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.667365 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"swift-proxy-config-data" Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.667992 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-swift-public-svc" Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.685391 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6486584599-dfqxj"] Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.829631 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dfc995c0-dbf3-45a0-b16e-3d86acabe454" path="/var/lib/kubelet/pods/dfc995c0-dbf3-45a0-b16e-3d86acabe454/volumes" Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.858213 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe5d2413-57a2-449d-9973-642e00e8c49f-combined-ca-bundle\") pod \"swift-proxy-6486584599-dfqxj\" (UID: \"fe5d2413-57a2-449d-9973-642e00e8c49f\") " pod="openstack/swift-proxy-6486584599-dfqxj" Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.858264 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wnk9j\" (UniqueName: \"kubernetes.io/projected/fe5d2413-57a2-449d-9973-642e00e8c49f-kube-api-access-wnk9j\") pod \"swift-proxy-6486584599-dfqxj\" (UID: \"fe5d2413-57a2-449d-9973-642e00e8c49f\") " pod="openstack/swift-proxy-6486584599-dfqxj" Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.858310 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fe5d2413-57a2-449d-9973-642e00e8c49f-run-httpd\") pod \"swift-proxy-6486584599-dfqxj\" (UID: \"fe5d2413-57a2-449d-9973-642e00e8c49f\") " pod="openstack/swift-proxy-6486584599-dfqxj" Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.858336 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe5d2413-57a2-449d-9973-642e00e8c49f-public-tls-certs\") pod \"swift-proxy-6486584599-dfqxj\" (UID: \"fe5d2413-57a2-449d-9973-642e00e8c49f\") " pod="openstack/swift-proxy-6486584599-dfqxj" Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.858370 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fe5d2413-57a2-449d-9973-642e00e8c49f-log-httpd\") pod \"swift-proxy-6486584599-dfqxj\" (UID: \"fe5d2413-57a2-449d-9973-642e00e8c49f\") " pod="openstack/swift-proxy-6486584599-dfqxj" Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.858414 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe5d2413-57a2-449d-9973-642e00e8c49f-internal-tls-certs\") pod \"swift-proxy-6486584599-dfqxj\" (UID: \"fe5d2413-57a2-449d-9973-642e00e8c49f\") " pod="openstack/swift-proxy-6486584599-dfqxj" Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.858468 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fe5d2413-57a2-449d-9973-642e00e8c49f-etc-swift\") pod \"swift-proxy-6486584599-dfqxj\" (UID: \"fe5d2413-57a2-449d-9973-642e00e8c49f\") " pod="openstack/swift-proxy-6486584599-dfqxj" Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.858513 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe5d2413-57a2-449d-9973-642e00e8c49f-config-data\") pod \"swift-proxy-6486584599-dfqxj\" (UID: \"fe5d2413-57a2-449d-9973-642e00e8c49f\") " pod="openstack/swift-proxy-6486584599-dfqxj" Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.960061 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fe5d2413-57a2-449d-9973-642e00e8c49f-etc-swift\") pod \"swift-proxy-6486584599-dfqxj\" (UID: \"fe5d2413-57a2-449d-9973-642e00e8c49f\") " pod="openstack/swift-proxy-6486584599-dfqxj" Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.960171 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe5d2413-57a2-449d-9973-642e00e8c49f-config-data\") pod \"swift-proxy-6486584599-dfqxj\" (UID: \"fe5d2413-57a2-449d-9973-642e00e8c49f\") " pod="openstack/swift-proxy-6486584599-dfqxj" Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.960254 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe5d2413-57a2-449d-9973-642e00e8c49f-combined-ca-bundle\") pod \"swift-proxy-6486584599-dfqxj\" (UID: \"fe5d2413-57a2-449d-9973-642e00e8c49f\") " pod="openstack/swift-proxy-6486584599-dfqxj" Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.960284 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wnk9j\" (UniqueName: \"kubernetes.io/projected/fe5d2413-57a2-449d-9973-642e00e8c49f-kube-api-access-wnk9j\") pod \"swift-proxy-6486584599-dfqxj\" (UID: \"fe5d2413-57a2-449d-9973-642e00e8c49f\") " pod="openstack/swift-proxy-6486584599-dfqxj" Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.960322 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fe5d2413-57a2-449d-9973-642e00e8c49f-run-httpd\") pod \"swift-proxy-6486584599-dfqxj\" (UID: \"fe5d2413-57a2-449d-9973-642e00e8c49f\") " pod="openstack/swift-proxy-6486584599-dfqxj" Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.960344 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe5d2413-57a2-449d-9973-642e00e8c49f-public-tls-certs\") pod \"swift-proxy-6486584599-dfqxj\" (UID: \"fe5d2413-57a2-449d-9973-642e00e8c49f\") " pod="openstack/swift-proxy-6486584599-dfqxj" Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.960374 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fe5d2413-57a2-449d-9973-642e00e8c49f-log-httpd\") pod \"swift-proxy-6486584599-dfqxj\" (UID: \"fe5d2413-57a2-449d-9973-642e00e8c49f\") " pod="openstack/swift-proxy-6486584599-dfqxj" Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.960414 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe5d2413-57a2-449d-9973-642e00e8c49f-internal-tls-certs\") pod \"swift-proxy-6486584599-dfqxj\" (UID: \"fe5d2413-57a2-449d-9973-642e00e8c49f\") " pod="openstack/swift-proxy-6486584599-dfqxj" Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.964344 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fe5d2413-57a2-449d-9973-642e00e8c49f-run-httpd\") pod \"swift-proxy-6486584599-dfqxj\" (UID: \"fe5d2413-57a2-449d-9973-642e00e8c49f\") " pod="openstack/swift-proxy-6486584599-dfqxj" Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.964428 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/fe5d2413-57a2-449d-9973-642e00e8c49f-log-httpd\") pod \"swift-proxy-6486584599-dfqxj\" (UID: \"fe5d2413-57a2-449d-9973-642e00e8c49f\") " pod="openstack/swift-proxy-6486584599-dfqxj" Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.968388 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-swift\" (UniqueName: \"kubernetes.io/projected/fe5d2413-57a2-449d-9973-642e00e8c49f-etc-swift\") pod \"swift-proxy-6486584599-dfqxj\" (UID: \"fe5d2413-57a2-449d-9973-642e00e8c49f\") " pod="openstack/swift-proxy-6486584599-dfqxj" Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.969204 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/fe5d2413-57a2-449d-9973-642e00e8c49f-config-data\") pod \"swift-proxy-6486584599-dfqxj\" (UID: \"fe5d2413-57a2-449d-9973-642e00e8c49f\") " pod="openstack/swift-proxy-6486584599-dfqxj" Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.969978 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe5d2413-57a2-449d-9973-642e00e8c49f-internal-tls-certs\") pod \"swift-proxy-6486584599-dfqxj\" (UID: \"fe5d2413-57a2-449d-9973-642e00e8c49f\") " pod="openstack/swift-proxy-6486584599-dfqxj" Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.970595 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/fe5d2413-57a2-449d-9973-642e00e8c49f-public-tls-certs\") pod \"swift-proxy-6486584599-dfqxj\" (UID: \"fe5d2413-57a2-449d-9973-642e00e8c49f\") " pod="openstack/swift-proxy-6486584599-dfqxj" Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.982365 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/fe5d2413-57a2-449d-9973-642e00e8c49f-combined-ca-bundle\") pod \"swift-proxy-6486584599-dfqxj\" (UID: \"fe5d2413-57a2-449d-9973-642e00e8c49f\") " pod="openstack/swift-proxy-6486584599-dfqxj" Oct 11 04:08:30 crc kubenswrapper[4967]: I1011 04:08:30.988124 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wnk9j\" (UniqueName: \"kubernetes.io/projected/fe5d2413-57a2-449d-9973-642e00e8c49f-kube-api-access-wnk9j\") pod \"swift-proxy-6486584599-dfqxj\" (UID: \"fe5d2413-57a2-449d-9973-642e00e8c49f\") " pod="openstack/swift-proxy-6486584599-dfqxj" Oct 11 04:08:31 crc kubenswrapper[4967]: I1011 04:08:31.276620 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/swift-proxy-6486584599-dfqxj" Oct 11 04:08:31 crc kubenswrapper[4967]: I1011 04:08:31.431280 4967 generic.go:334] "Generic (PLEG): container finished" podID="1680354a-b333-4e21-9748-0062d78a479c" containerID="433cef11bc0a2150c9ca5cf11665083f8de67f71d9f8a86c58ef07e0b0409f7e" exitCode=0 Oct 11 04:08:31 crc kubenswrapper[4967]: I1011 04:08:31.431488 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-6bhhf" event={"ID":"1680354a-b333-4e21-9748-0062d78a479c","Type":"ContainerDied","Data":"433cef11bc0a2150c9ca5cf11665083f8de67f71d9f8a86c58ef07e0b0409f7e"} Oct 11 04:08:31 crc kubenswrapper[4967]: I1011 04:08:31.914434 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/swift-proxy-6486584599-dfqxj"] Oct 11 04:08:31 crc kubenswrapper[4967]: W1011 04:08:31.919826 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podfe5d2413_57a2_449d_9973_642e00e8c49f.slice/crio-915603e39ff4b3eaed54ff11e45c277c3182ea228c3539d834285662fcb69314 WatchSource:0}: Error finding container 915603e39ff4b3eaed54ff11e45c277c3182ea228c3539d834285662fcb69314: Status 404 returned error can't find the container with id 915603e39ff4b3eaed54ff11e45c277c3182ea228c3539d834285662fcb69314 Oct 11 04:08:32 crc kubenswrapper[4967]: I1011 04:08:32.443283 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6486584599-dfqxj" event={"ID":"fe5d2413-57a2-449d-9973-642e00e8c49f","Type":"ContainerStarted","Data":"9fa2f4760f0bce0a5a108a15553becb2ca25fc703e6cb07aa7c12580cbf916b4"} Oct 11 04:08:32 crc kubenswrapper[4967]: I1011 04:08:32.443634 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6486584599-dfqxj" event={"ID":"fe5d2413-57a2-449d-9973-642e00e8c49f","Type":"ContainerStarted","Data":"9014bc0ea24a4f1b8a2f53de00d400ea418d2aefa4def13d202eda87115287a6"} Oct 11 04:08:32 crc kubenswrapper[4967]: I1011 04:08:32.443649 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/swift-proxy-6486584599-dfqxj" event={"ID":"fe5d2413-57a2-449d-9973-642e00e8c49f","Type":"ContainerStarted","Data":"915603e39ff4b3eaed54ff11e45c277c3182ea228c3539d834285662fcb69314"} Oct 11 04:08:32 crc kubenswrapper[4967]: I1011 04:08:32.443667 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6486584599-dfqxj" Oct 11 04:08:32 crc kubenswrapper[4967]: I1011 04:08:32.443680 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/swift-proxy-6486584599-dfqxj" Oct 11 04:08:32 crc kubenswrapper[4967]: I1011 04:08:32.470690 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/swift-proxy-6486584599-dfqxj" podStartSLOduration=2.470668483 podStartE2EDuration="2.470668483s" podCreationTimestamp="2025-10-11 04:08:30 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:08:32.470059262 +0000 UTC m=+1040.433268225" watchObservedRunningTime="2025-10-11 04:08:32.470668483 +0000 UTC m=+1040.433877416" Oct 11 04:08:32 crc kubenswrapper[4967]: I1011 04:08:32.737122 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-6bhhf" Oct 11 04:08:32 crc kubenswrapper[4967]: I1011 04:08:32.895625 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1680354a-b333-4e21-9748-0062d78a479c-config-data\") pod \"1680354a-b333-4e21-9748-0062d78a479c\" (UID: \"1680354a-b333-4e21-9748-0062d78a479c\") " Oct 11 04:08:32 crc kubenswrapper[4967]: I1011 04:08:32.895930 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1680354a-b333-4e21-9748-0062d78a479c-etc-machine-id\") pod \"1680354a-b333-4e21-9748-0062d78a479c\" (UID: \"1680354a-b333-4e21-9748-0062d78a479c\") " Oct 11 04:08:32 crc kubenswrapper[4967]: I1011 04:08:32.895960 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1680354a-b333-4e21-9748-0062d78a479c-combined-ca-bundle\") pod \"1680354a-b333-4e21-9748-0062d78a479c\" (UID: \"1680354a-b333-4e21-9748-0062d78a479c\") " Oct 11 04:08:32 crc kubenswrapper[4967]: I1011 04:08:32.895987 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1680354a-b333-4e21-9748-0062d78a479c-db-sync-config-data\") pod \"1680354a-b333-4e21-9748-0062d78a479c\" (UID: \"1680354a-b333-4e21-9748-0062d78a479c\") " Oct 11 04:08:32 crc kubenswrapper[4967]: I1011 04:08:32.896014 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dq4ts\" (UniqueName: \"kubernetes.io/projected/1680354a-b333-4e21-9748-0062d78a479c-kube-api-access-dq4ts\") pod \"1680354a-b333-4e21-9748-0062d78a479c\" (UID: \"1680354a-b333-4e21-9748-0062d78a479c\") " Oct 11 04:08:32 crc kubenswrapper[4967]: I1011 04:08:32.896051 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1680354a-b333-4e21-9748-0062d78a479c-scripts\") pod \"1680354a-b333-4e21-9748-0062d78a479c\" (UID: \"1680354a-b333-4e21-9748-0062d78a479c\") " Oct 11 04:08:32 crc kubenswrapper[4967]: I1011 04:08:32.897200 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/1680354a-b333-4e21-9748-0062d78a479c-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "1680354a-b333-4e21-9748-0062d78a479c" (UID: "1680354a-b333-4e21-9748-0062d78a479c"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 04:08:32 crc kubenswrapper[4967]: I1011 04:08:32.901905 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1680354a-b333-4e21-9748-0062d78a479c-db-sync-config-data" (OuterVolumeSpecName: "db-sync-config-data") pod "1680354a-b333-4e21-9748-0062d78a479c" (UID: "1680354a-b333-4e21-9748-0062d78a479c"). InnerVolumeSpecName "db-sync-config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:32 crc kubenswrapper[4967]: I1011 04:08:32.902487 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1680354a-b333-4e21-9748-0062d78a479c-kube-api-access-dq4ts" (OuterVolumeSpecName: "kube-api-access-dq4ts") pod "1680354a-b333-4e21-9748-0062d78a479c" (UID: "1680354a-b333-4e21-9748-0062d78a479c"). InnerVolumeSpecName "kube-api-access-dq4ts". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:08:32 crc kubenswrapper[4967]: I1011 04:08:32.902674 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1680354a-b333-4e21-9748-0062d78a479c-scripts" (OuterVolumeSpecName: "scripts") pod "1680354a-b333-4e21-9748-0062d78a479c" (UID: "1680354a-b333-4e21-9748-0062d78a479c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:32 crc kubenswrapper[4967]: I1011 04:08:32.924333 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1680354a-b333-4e21-9748-0062d78a479c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1680354a-b333-4e21-9748-0062d78a479c" (UID: "1680354a-b333-4e21-9748-0062d78a479c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:32 crc kubenswrapper[4967]: I1011 04:08:32.970477 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1680354a-b333-4e21-9748-0062d78a479c-config-data" (OuterVolumeSpecName: "config-data") pod "1680354a-b333-4e21-9748-0062d78a479c" (UID: "1680354a-b333-4e21-9748-0062d78a479c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:32 crc kubenswrapper[4967]: I1011 04:08:32.998121 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1680354a-b333-4e21-9748-0062d78a479c-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:32 crc kubenswrapper[4967]: I1011 04:08:32.998153 4967 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/1680354a-b333-4e21-9748-0062d78a479c-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:32 crc kubenswrapper[4967]: I1011 04:08:32.998163 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1680354a-b333-4e21-9748-0062d78a479c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:32 crc kubenswrapper[4967]: I1011 04:08:32.998172 4967 reconciler_common.go:293] "Volume detached for volume \"db-sync-config-data\" (UniqueName: \"kubernetes.io/secret/1680354a-b333-4e21-9748-0062d78a479c-db-sync-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:32 crc kubenswrapper[4967]: I1011 04:08:32.998181 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dq4ts\" (UniqueName: \"kubernetes.io/projected/1680354a-b333-4e21-9748-0062d78a479c-kube-api-access-dq4ts\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:32 crc kubenswrapper[4967]: I1011 04:08:32.998188 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1680354a-b333-4e21-9748-0062d78a479c-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.265726 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.266065 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9d5e1373-95a7-4ec3-9316-c24cc9ee8833" containerName="ceilometer-central-agent" containerID="cri-o://dae619182846b7310ad63ec9ee260e7711654bf0e8b68976b12a1d8fdcad126c" gracePeriod=30 Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.267037 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9d5e1373-95a7-4ec3-9316-c24cc9ee8833" containerName="proxy-httpd" containerID="cri-o://f64b25630366a663ac416ada48352e991a15a0213371220d837e72ba513411e1" gracePeriod=30 Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.267187 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9d5e1373-95a7-4ec3-9316-c24cc9ee8833" containerName="ceilometer-notification-agent" containerID="cri-o://c4d97530d947be64f12582efe3328e97cc2041a7c4b16e0fd87b9768a2dae2f0" gracePeriod=30 Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.267239 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="9d5e1373-95a7-4ec3-9316-c24cc9ee8833" containerName="sg-core" containerID="cri-o://7236195b1efa93e02f72597cbd63168d14fddc957d15a80644b2c0d778880cb6" gracePeriod=30 Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.461679 4967 generic.go:334] "Generic (PLEG): container finished" podID="9d5e1373-95a7-4ec3-9316-c24cc9ee8833" containerID="7236195b1efa93e02f72597cbd63168d14fddc957d15a80644b2c0d778880cb6" exitCode=2 Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.461733 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9d5e1373-95a7-4ec3-9316-c24cc9ee8833","Type":"ContainerDied","Data":"7236195b1efa93e02f72597cbd63168d14fddc957d15a80644b2c0d778880cb6"} Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.482856 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-db-sync-6bhhf" event={"ID":"1680354a-b333-4e21-9748-0062d78a479c","Type":"ContainerDied","Data":"7d6e1e33ad71585b4c706ec5ff964a445a52f317855abab60703336fc3333040"} Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.482904 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7d6e1e33ad71585b4c706ec5ff964a445a52f317855abab60703336fc3333040" Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.482979 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-db-sync-6bhhf" Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.745033 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 11 04:08:33 crc kubenswrapper[4967]: E1011 04:08:33.745676 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1680354a-b333-4e21-9748-0062d78a479c" containerName="cinder-db-sync" Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.745687 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="1680354a-b333-4e21-9748-0062d78a479c" containerName="cinder-db-sync" Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.745859 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="1680354a-b333-4e21-9748-0062d78a479c" containerName="cinder-db-sync" Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.746766 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.750951 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.751498 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scripts" Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.751739 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-config-data" Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.751848 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-cinder-dockercfg-24g9r" Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.757536 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.801186 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-vhtnn"] Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.802704 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-vhtnn" Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.843537 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-vhtnn"] Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.914833 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bdd15be-7524-4447-bab5-43c0a6b40cf0-config-data\") pod \"cinder-scheduler-0\" (UID: \"4bdd15be-7524-4447-bab5-43c0a6b40cf0\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.914872 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-svs7q\" (UniqueName: \"kubernetes.io/projected/4bdd15be-7524-4447-bab5-43c0a6b40cf0-kube-api-access-svs7q\") pod \"cinder-scheduler-0\" (UID: \"4bdd15be-7524-4447-bab5-43c0a6b40cf0\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.914901 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-vhtnn\" (UID: \"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291\") " pod="openstack/dnsmasq-dns-6578955fd5-vhtnn" Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.914927 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-dns-svc\") pod \"dnsmasq-dns-6578955fd5-vhtnn\" (UID: \"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291\") " pod="openstack/dnsmasq-dns-6578955fd5-vhtnn" Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.914964 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z56rs\" (UniqueName: \"kubernetes.io/projected/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-kube-api-access-z56rs\") pod \"dnsmasq-dns-6578955fd5-vhtnn\" (UID: \"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291\") " pod="openstack/dnsmasq-dns-6578955fd5-vhtnn" Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.915025 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bdd15be-7524-4447-bab5-43c0a6b40cf0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4bdd15be-7524-4447-bab5-43c0a6b40cf0\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.915085 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4bdd15be-7524-4447-bab5-43c0a6b40cf0-scripts\") pod \"cinder-scheduler-0\" (UID: \"4bdd15be-7524-4447-bab5-43c0a6b40cf0\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.915104 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-vhtnn\" (UID: \"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291\") " pod="openstack/dnsmasq-dns-6578955fd5-vhtnn" Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.915120 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-vhtnn\" (UID: \"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291\") " pod="openstack/dnsmasq-dns-6578955fd5-vhtnn" Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.915135 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-config\") pod \"dnsmasq-dns-6578955fd5-vhtnn\" (UID: \"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291\") " pod="openstack/dnsmasq-dns-6578955fd5-vhtnn" Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.915160 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4bdd15be-7524-4447-bab5-43c0a6b40cf0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4bdd15be-7524-4447-bab5-43c0a6b40cf0\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.915198 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4bdd15be-7524-4447-bab5-43c0a6b40cf0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4bdd15be-7524-4447-bab5-43c0a6b40cf0\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.983393 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.985089 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 11 04:08:33 crc kubenswrapper[4967]: I1011 04:08:33.992818 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.016151 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.017959 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-vhtnn\" (UID: \"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291\") " pod="openstack/dnsmasq-dns-6578955fd5-vhtnn" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.017995 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-config\") pod \"dnsmasq-dns-6578955fd5-vhtnn\" (UID: \"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291\") " pod="openstack/dnsmasq-dns-6578955fd5-vhtnn" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.018032 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4bdd15be-7524-4447-bab5-43c0a6b40cf0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4bdd15be-7524-4447-bab5-43c0a6b40cf0\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.018082 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4bdd15be-7524-4447-bab5-43c0a6b40cf0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4bdd15be-7524-4447-bab5-43c0a6b40cf0\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.018106 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bdd15be-7524-4447-bab5-43c0a6b40cf0-config-data\") pod \"cinder-scheduler-0\" (UID: \"4bdd15be-7524-4447-bab5-43c0a6b40cf0\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.018125 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-svs7q\" (UniqueName: \"kubernetes.io/projected/4bdd15be-7524-4447-bab5-43c0a6b40cf0-kube-api-access-svs7q\") pod \"cinder-scheduler-0\" (UID: \"4bdd15be-7524-4447-bab5-43c0a6b40cf0\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.018153 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-vhtnn\" (UID: \"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291\") " pod="openstack/dnsmasq-dns-6578955fd5-vhtnn" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.018173 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-dns-svc\") pod \"dnsmasq-dns-6578955fd5-vhtnn\" (UID: \"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291\") " pod="openstack/dnsmasq-dns-6578955fd5-vhtnn" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.018203 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z56rs\" (UniqueName: \"kubernetes.io/projected/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-kube-api-access-z56rs\") pod \"dnsmasq-dns-6578955fd5-vhtnn\" (UID: \"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291\") " pod="openstack/dnsmasq-dns-6578955fd5-vhtnn" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.018253 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bdd15be-7524-4447-bab5-43c0a6b40cf0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4bdd15be-7524-4447-bab5-43c0a6b40cf0\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.018297 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4bdd15be-7524-4447-bab5-43c0a6b40cf0-scripts\") pod \"cinder-scheduler-0\" (UID: \"4bdd15be-7524-4447-bab5-43c0a6b40cf0\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.018316 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-vhtnn\" (UID: \"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291\") " pod="openstack/dnsmasq-dns-6578955fd5-vhtnn" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.019117 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-ovsdbserver-sb\") pod \"dnsmasq-dns-6578955fd5-vhtnn\" (UID: \"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291\") " pod="openstack/dnsmasq-dns-6578955fd5-vhtnn" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.019626 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-ovsdbserver-nb\") pod \"dnsmasq-dns-6578955fd5-vhtnn\" (UID: \"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291\") " pod="openstack/dnsmasq-dns-6578955fd5-vhtnn" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.020135 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-config\") pod \"dnsmasq-dns-6578955fd5-vhtnn\" (UID: \"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291\") " pod="openstack/dnsmasq-dns-6578955fd5-vhtnn" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.020651 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-dns-svc\") pod \"dnsmasq-dns-6578955fd5-vhtnn\" (UID: \"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291\") " pod="openstack/dnsmasq-dns-6578955fd5-vhtnn" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.020652 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-dns-swift-storage-0\") pod \"dnsmasq-dns-6578955fd5-vhtnn\" (UID: \"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291\") " pod="openstack/dnsmasq-dns-6578955fd5-vhtnn" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.020698 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4bdd15be-7524-4447-bab5-43c0a6b40cf0-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"4bdd15be-7524-4447-bab5-43c0a6b40cf0\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.032857 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4bdd15be-7524-4447-bab5-43c0a6b40cf0-scripts\") pod \"cinder-scheduler-0\" (UID: \"4bdd15be-7524-4447-bab5-43c0a6b40cf0\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.044318 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bdd15be-7524-4447-bab5-43c0a6b40cf0-config-data\") pod \"cinder-scheduler-0\" (UID: \"4bdd15be-7524-4447-bab5-43c0a6b40cf0\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.051530 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bdd15be-7524-4447-bab5-43c0a6b40cf0-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"4bdd15be-7524-4447-bab5-43c0a6b40cf0\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.055440 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-svs7q\" (UniqueName: \"kubernetes.io/projected/4bdd15be-7524-4447-bab5-43c0a6b40cf0-kube-api-access-svs7q\") pod \"cinder-scheduler-0\" (UID: \"4bdd15be-7524-4447-bab5-43c0a6b40cf0\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.070779 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4bdd15be-7524-4447-bab5-43c0a6b40cf0-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"4bdd15be-7524-4447-bab5-43c0a6b40cf0\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.081710 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z56rs\" (UniqueName: \"kubernetes.io/projected/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-kube-api-access-z56rs\") pod \"dnsmasq-dns-6578955fd5-vhtnn\" (UID: \"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291\") " pod="openstack/dnsmasq-dns-6578955fd5-vhtnn" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.088772 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.121010 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-config-data\") pod \"cinder-api-0\" (UID: \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\") " pod="openstack/cinder-api-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.121057 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-logs\") pod \"cinder-api-0\" (UID: \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\") " pod="openstack/cinder-api-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.127334 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\") " pod="openstack/cinder-api-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.127413 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lr5wf\" (UniqueName: \"kubernetes.io/projected/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-kube-api-access-lr5wf\") pod \"cinder-api-0\" (UID: \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\") " pod="openstack/cinder-api-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.127444 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\") " pod="openstack/cinder-api-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.127514 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-config-data-custom\") pod \"cinder-api-0\" (UID: \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\") " pod="openstack/cinder-api-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.127644 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-scripts\") pod \"cinder-api-0\" (UID: \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\") " pod="openstack/cinder-api-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.168974 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-vhtnn" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.229037 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-config-data\") pod \"cinder-api-0\" (UID: \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\") " pod="openstack/cinder-api-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.229350 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-logs\") pod \"cinder-api-0\" (UID: \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\") " pod="openstack/cinder-api-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.229413 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\") " pod="openstack/cinder-api-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.229435 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lr5wf\" (UniqueName: \"kubernetes.io/projected/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-kube-api-access-lr5wf\") pod \"cinder-api-0\" (UID: \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\") " pod="openstack/cinder-api-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.229453 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\") " pod="openstack/cinder-api-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.229486 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-config-data-custom\") pod \"cinder-api-0\" (UID: \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\") " pod="openstack/cinder-api-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.229542 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-scripts\") pod \"cinder-api-0\" (UID: \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\") " pod="openstack/cinder-api-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.232731 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-etc-machine-id\") pod \"cinder-api-0\" (UID: \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\") " pod="openstack/cinder-api-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.233560 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-logs\") pod \"cinder-api-0\" (UID: \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\") " pod="openstack/cinder-api-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.235743 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-scripts\") pod \"cinder-api-0\" (UID: \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\") " pod="openstack/cinder-api-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.239404 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-config-data-custom\") pod \"cinder-api-0\" (UID: \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\") " pod="openstack/cinder-api-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.242800 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\") " pod="openstack/cinder-api-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.242889 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-config-data\") pod \"cinder-api-0\" (UID: \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\") " pod="openstack/cinder-api-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.258449 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lr5wf\" (UniqueName: \"kubernetes.io/projected/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-kube-api-access-lr5wf\") pod \"cinder-api-0\" (UID: \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\") " pod="openstack/cinder-api-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.499839 4967 generic.go:334] "Generic (PLEG): container finished" podID="9d5e1373-95a7-4ec3-9316-c24cc9ee8833" containerID="f64b25630366a663ac416ada48352e991a15a0213371220d837e72ba513411e1" exitCode=0 Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.499866 4967 generic.go:334] "Generic (PLEG): container finished" podID="9d5e1373-95a7-4ec3-9316-c24cc9ee8833" containerID="c4d97530d947be64f12582efe3328e97cc2041a7c4b16e0fd87b9768a2dae2f0" exitCode=0 Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.499875 4967 generic.go:334] "Generic (PLEG): container finished" podID="9d5e1373-95a7-4ec3-9316-c24cc9ee8833" containerID="dae619182846b7310ad63ec9ee260e7711654bf0e8b68976b12a1d8fdcad126c" exitCode=0 Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.499894 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9d5e1373-95a7-4ec3-9316-c24cc9ee8833","Type":"ContainerDied","Data":"f64b25630366a663ac416ada48352e991a15a0213371220d837e72ba513411e1"} Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.499918 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9d5e1373-95a7-4ec3-9316-c24cc9ee8833","Type":"ContainerDied","Data":"c4d97530d947be64f12582efe3328e97cc2041a7c4b16e0fd87b9768a2dae2f0"} Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.499928 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9d5e1373-95a7-4ec3-9316-c24cc9ee8833","Type":"ContainerDied","Data":"dae619182846b7310ad63ec9ee260e7711654bf0e8b68976b12a1d8fdcad126c"} Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.523360 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.644533 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.779581 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-vhtnn"] Oct 11 04:08:34 crc kubenswrapper[4967]: I1011 04:08:34.893852 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.065673 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-sg-core-conf-yaml\") pod \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\" (UID: \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\") " Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.065750 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-run-httpd\") pod \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\" (UID: \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\") " Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.065869 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-config-data\") pod \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\" (UID: \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\") " Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.065915 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-combined-ca-bundle\") pod \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\" (UID: \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\") " Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.065968 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-log-httpd\") pod \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\" (UID: \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\") " Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.065990 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-scripts\") pod \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\" (UID: \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\") " Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.066064 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-slqbp\" (UniqueName: \"kubernetes.io/projected/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-kube-api-access-slqbp\") pod \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\" (UID: \"9d5e1373-95a7-4ec3-9316-c24cc9ee8833\") " Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.076405 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "9d5e1373-95a7-4ec3-9316-c24cc9ee8833" (UID: "9d5e1373-95a7-4ec3-9316-c24cc9ee8833"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.076586 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "9d5e1373-95a7-4ec3-9316-c24cc9ee8833" (UID: "9d5e1373-95a7-4ec3-9316-c24cc9ee8833"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.080816 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-kube-api-access-slqbp" (OuterVolumeSpecName: "kube-api-access-slqbp") pod "9d5e1373-95a7-4ec3-9316-c24cc9ee8833" (UID: "9d5e1373-95a7-4ec3-9316-c24cc9ee8833"). InnerVolumeSpecName "kube-api-access-slqbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.103340 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-scripts" (OuterVolumeSpecName: "scripts") pod "9d5e1373-95a7-4ec3-9316-c24cc9ee8833" (UID: "9d5e1373-95a7-4ec3-9316-c24cc9ee8833"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.124987 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "9d5e1373-95a7-4ec3-9316-c24cc9ee8833" (UID: "9d5e1373-95a7-4ec3-9316-c24cc9ee8833"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.162127 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.167816 4967 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.167847 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.167856 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-slqbp\" (UniqueName: \"kubernetes.io/projected/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-kube-api-access-slqbp\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.167866 4967 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.167875 4967 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.307790 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9d5e1373-95a7-4ec3-9316-c24cc9ee8833" (UID: "9d5e1373-95a7-4ec3-9316-c24cc9ee8833"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.319770 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-config-data" (OuterVolumeSpecName: "config-data") pod "9d5e1373-95a7-4ec3-9316-c24cc9ee8833" (UID: "9d5e1373-95a7-4ec3-9316-c24cc9ee8833"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.372434 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.372466 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9d5e1373-95a7-4ec3-9316-c24cc9ee8833-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.517471 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4bdd15be-7524-4447-bab5-43c0a6b40cf0","Type":"ContainerStarted","Data":"5eb186429e6c0f24e596aec4d6e354562066fb6d2708ca4aa4ac9a019e2da3e0"} Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.529862 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-84674df998-htlks" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.563809 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6633fd1a-cd77-429d-b24f-33cb1bcba3f3","Type":"ContainerStarted","Data":"931caf0e691edc8d915cc559cf6c851833dbc13584ef8af86015734bad24ea73"} Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.574462 4967 generic.go:334] "Generic (PLEG): container finished" podID="60a1ad86-dc2a-46a3-8c63-6e48d3a2d291" containerID="da1a30c39594c793da68e5ba312b7e700b454828fae3142a32c3eddfaaa5e815" exitCode=0 Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.574566 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-vhtnn" event={"ID":"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291","Type":"ContainerDied","Data":"da1a30c39594c793da68e5ba312b7e700b454828fae3142a32c3eddfaaa5e815"} Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.574621 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-vhtnn" event={"ID":"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291","Type":"ContainerStarted","Data":"1e99736afe34e37994f390afffddc363850018f5aaeb843222ffc2ba17efac02"} Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.633238 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/barbican-api-84674df998-htlks" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.635094 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"9d5e1373-95a7-4ec3-9316-c24cc9ee8833","Type":"ContainerDied","Data":"1b1309c358fc7af4082c92f069e845f6d0876cf6f208f8c161141c49291ecd94"} Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.635163 4967 scope.go:117] "RemoveContainer" containerID="f64b25630366a663ac416ada48352e991a15a0213371220d837e72ba513411e1" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.635358 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.693211 4967 scope.go:117] "RemoveContainer" containerID="7236195b1efa93e02f72597cbd63168d14fddc957d15a80644b2c0d778880cb6" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.698426 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6b6c6c8584-4qwkb"] Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.698898 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6b6c6c8584-4qwkb" podUID="9c6c522c-bd96-42f6-b0ae-27d55d3121c4" containerName="barbican-api-log" containerID="cri-o://2713f1c59d64482f421e2cd9da89a876d483393a3a347f05ed054178dad43dfa" gracePeriod=30 Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.699027 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/barbican-api-6b6c6c8584-4qwkb" podUID="9c6c522c-bd96-42f6-b0ae-27d55d3121c4" containerName="barbican-api" containerID="cri-o://fab128d00e2d02acf1adf0f691a5c9d7a4302c62e8039f94ba4a887ed5d74b69" gracePeriod=30 Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.726134 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.756999 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.777961 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:08:35 crc kubenswrapper[4967]: E1011 04:08:35.778311 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d5e1373-95a7-4ec3-9316-c24cc9ee8833" containerName="ceilometer-notification-agent" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.778324 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d5e1373-95a7-4ec3-9316-c24cc9ee8833" containerName="ceilometer-notification-agent" Oct 11 04:08:35 crc kubenswrapper[4967]: E1011 04:08:35.778342 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d5e1373-95a7-4ec3-9316-c24cc9ee8833" containerName="ceilometer-central-agent" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.778349 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d5e1373-95a7-4ec3-9316-c24cc9ee8833" containerName="ceilometer-central-agent" Oct 11 04:08:35 crc kubenswrapper[4967]: E1011 04:08:35.778359 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d5e1373-95a7-4ec3-9316-c24cc9ee8833" containerName="proxy-httpd" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.778365 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d5e1373-95a7-4ec3-9316-c24cc9ee8833" containerName="proxy-httpd" Oct 11 04:08:35 crc kubenswrapper[4967]: E1011 04:08:35.778387 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9d5e1373-95a7-4ec3-9316-c24cc9ee8833" containerName="sg-core" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.778392 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="9d5e1373-95a7-4ec3-9316-c24cc9ee8833" containerName="sg-core" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.778538 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d5e1373-95a7-4ec3-9316-c24cc9ee8833" containerName="proxy-httpd" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.778557 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d5e1373-95a7-4ec3-9316-c24cc9ee8833" containerName="ceilometer-central-agent" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.778567 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d5e1373-95a7-4ec3-9316-c24cc9ee8833" containerName="sg-core" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.778580 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="9d5e1373-95a7-4ec3-9316-c24cc9ee8833" containerName="ceilometer-notification-agent" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.780017 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.791776 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.791939 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.824779 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.836255 4967 scope.go:117] "RemoveContainer" containerID="c4d97530d947be64f12582efe3328e97cc2041a7c4b16e0fd87b9768a2dae2f0" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.891812 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8e8fb63d-44cd-427e-8324-743d83fc8416-run-httpd\") pod \"ceilometer-0\" (UID: \"8e8fb63d-44cd-427e-8324-743d83fc8416\") " pod="openstack/ceilometer-0" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.891870 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bvstw\" (UniqueName: \"kubernetes.io/projected/8e8fb63d-44cd-427e-8324-743d83fc8416-kube-api-access-bvstw\") pod \"ceilometer-0\" (UID: \"8e8fb63d-44cd-427e-8324-743d83fc8416\") " pod="openstack/ceilometer-0" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.891914 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8e8fb63d-44cd-427e-8324-743d83fc8416-log-httpd\") pod \"ceilometer-0\" (UID: \"8e8fb63d-44cd-427e-8324-743d83fc8416\") " pod="openstack/ceilometer-0" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.891945 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e8fb63d-44cd-427e-8324-743d83fc8416-scripts\") pod \"ceilometer-0\" (UID: \"8e8fb63d-44cd-427e-8324-743d83fc8416\") " pod="openstack/ceilometer-0" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.891971 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8e8fb63d-44cd-427e-8324-743d83fc8416-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8e8fb63d-44cd-427e-8324-743d83fc8416\") " pod="openstack/ceilometer-0" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.892011 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e8fb63d-44cd-427e-8324-743d83fc8416-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8e8fb63d-44cd-427e-8324-743d83fc8416\") " pod="openstack/ceilometer-0" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.892029 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e8fb63d-44cd-427e-8324-743d83fc8416-config-data\") pod \"ceilometer-0\" (UID: \"8e8fb63d-44cd-427e-8324-743d83fc8416\") " pod="openstack/ceilometer-0" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.941803 4967 scope.go:117] "RemoveContainer" containerID="dae619182846b7310ad63ec9ee260e7711654bf0e8b68976b12a1d8fdcad126c" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.993548 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bvstw\" (UniqueName: \"kubernetes.io/projected/8e8fb63d-44cd-427e-8324-743d83fc8416-kube-api-access-bvstw\") pod \"ceilometer-0\" (UID: \"8e8fb63d-44cd-427e-8324-743d83fc8416\") " pod="openstack/ceilometer-0" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.994097 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8e8fb63d-44cd-427e-8324-743d83fc8416-log-httpd\") pod \"ceilometer-0\" (UID: \"8e8fb63d-44cd-427e-8324-743d83fc8416\") " pod="openstack/ceilometer-0" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.994173 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e8fb63d-44cd-427e-8324-743d83fc8416-scripts\") pod \"ceilometer-0\" (UID: \"8e8fb63d-44cd-427e-8324-743d83fc8416\") " pod="openstack/ceilometer-0" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.994218 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8e8fb63d-44cd-427e-8324-743d83fc8416-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8e8fb63d-44cd-427e-8324-743d83fc8416\") " pod="openstack/ceilometer-0" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.994368 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e8fb63d-44cd-427e-8324-743d83fc8416-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8e8fb63d-44cd-427e-8324-743d83fc8416\") " pod="openstack/ceilometer-0" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.994398 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e8fb63d-44cd-427e-8324-743d83fc8416-config-data\") pod \"ceilometer-0\" (UID: \"8e8fb63d-44cd-427e-8324-743d83fc8416\") " pod="openstack/ceilometer-0" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.994504 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8e8fb63d-44cd-427e-8324-743d83fc8416-run-httpd\") pod \"ceilometer-0\" (UID: \"8e8fb63d-44cd-427e-8324-743d83fc8416\") " pod="openstack/ceilometer-0" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.995049 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8e8fb63d-44cd-427e-8324-743d83fc8416-run-httpd\") pod \"ceilometer-0\" (UID: \"8e8fb63d-44cd-427e-8324-743d83fc8416\") " pod="openstack/ceilometer-0" Oct 11 04:08:35 crc kubenswrapper[4967]: I1011 04:08:35.995811 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8e8fb63d-44cd-427e-8324-743d83fc8416-log-httpd\") pod \"ceilometer-0\" (UID: \"8e8fb63d-44cd-427e-8324-743d83fc8416\") " pod="openstack/ceilometer-0" Oct 11 04:08:36 crc kubenswrapper[4967]: I1011 04:08:35.998555 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8e8fb63d-44cd-427e-8324-743d83fc8416-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"8e8fb63d-44cd-427e-8324-743d83fc8416\") " pod="openstack/ceilometer-0" Oct 11 04:08:36 crc kubenswrapper[4967]: I1011 04:08:35.998727 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e8fb63d-44cd-427e-8324-743d83fc8416-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"8e8fb63d-44cd-427e-8324-743d83fc8416\") " pod="openstack/ceilometer-0" Oct 11 04:08:36 crc kubenswrapper[4967]: I1011 04:08:36.000897 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e8fb63d-44cd-427e-8324-743d83fc8416-scripts\") pod \"ceilometer-0\" (UID: \"8e8fb63d-44cd-427e-8324-743d83fc8416\") " pod="openstack/ceilometer-0" Oct 11 04:08:36 crc kubenswrapper[4967]: I1011 04:08:36.004887 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e8fb63d-44cd-427e-8324-743d83fc8416-config-data\") pod \"ceilometer-0\" (UID: \"8e8fb63d-44cd-427e-8324-743d83fc8416\") " pod="openstack/ceilometer-0" Oct 11 04:08:36 crc kubenswrapper[4967]: I1011 04:08:36.017601 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bvstw\" (UniqueName: \"kubernetes.io/projected/8e8fb63d-44cd-427e-8324-743d83fc8416-kube-api-access-bvstw\") pod \"ceilometer-0\" (UID: \"8e8fb63d-44cd-427e-8324-743d83fc8416\") " pod="openstack/ceilometer-0" Oct 11 04:08:36 crc kubenswrapper[4967]: I1011 04:08:36.146972 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:08:36 crc kubenswrapper[4967]: I1011 04:08:36.373870 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 11 04:08:36 crc kubenswrapper[4967]: I1011 04:08:36.646447 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6633fd1a-cd77-429d-b24f-33cb1bcba3f3","Type":"ContainerStarted","Data":"0e9d2174e592c5fa8e61611c5b8336748ff5343d122e6c90769747d418b2208d"} Oct 11 04:08:36 crc kubenswrapper[4967]: I1011 04:08:36.648717 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-vhtnn" event={"ID":"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291","Type":"ContainerStarted","Data":"f98f49cc939cb150c97d159f63d67e4ee4bfc1f905784b0d426a80a71c4386a3"} Oct 11 04:08:36 crc kubenswrapper[4967]: I1011 04:08:36.651700 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-6578955fd5-vhtnn" Oct 11 04:08:36 crc kubenswrapper[4967]: I1011 04:08:36.662296 4967 generic.go:334] "Generic (PLEG): container finished" podID="9c6c522c-bd96-42f6-b0ae-27d55d3121c4" containerID="2713f1c59d64482f421e2cd9da89a876d483393a3a347f05ed054178dad43dfa" exitCode=143 Oct 11 04:08:36 crc kubenswrapper[4967]: I1011 04:08:36.662450 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b6c6c8584-4qwkb" event={"ID":"9c6c522c-bd96-42f6-b0ae-27d55d3121c4","Type":"ContainerDied","Data":"2713f1c59d64482f421e2cd9da89a876d483393a3a347f05ed054178dad43dfa"} Oct 11 04:08:36 crc kubenswrapper[4967]: I1011 04:08:36.667701 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-6578955fd5-vhtnn" podStartSLOduration=3.667689378 podStartE2EDuration="3.667689378s" podCreationTimestamp="2025-10-11 04:08:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:08:36.665583441 +0000 UTC m=+1044.628792374" watchObservedRunningTime="2025-10-11 04:08:36.667689378 +0000 UTC m=+1044.630898311" Oct 11 04:08:36 crc kubenswrapper[4967]: I1011 04:08:36.835021 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d5e1373-95a7-4ec3-9316-c24cc9ee8833" path="/var/lib/kubelet/pods/9d5e1373-95a7-4ec3-9316-c24cc9ee8833/volumes" Oct 11 04:08:36 crc kubenswrapper[4967]: I1011 04:08:36.837401 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:08:37 crc kubenswrapper[4967]: I1011 04:08:37.059023 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:08:37 crc kubenswrapper[4967]: I1011 04:08:37.692390 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4bdd15be-7524-4447-bab5-43c0a6b40cf0","Type":"ContainerStarted","Data":"89575635b09e4fd75c0a08808d714ea4dee6937997a73292f38f9e45e2e88460"} Oct 11 04:08:38 crc kubenswrapper[4967]: I1011 04:08:38.720951 4967 generic.go:334] "Generic (PLEG): container finished" podID="1d249b36-580e-424d-8f36-e366c932411c" containerID="02bf66457fc098fe15c42a14af57eb113a56800408703ed35ed3a2b4f07ab74e" exitCode=0 Oct 11 04:08:38 crc kubenswrapper[4967]: I1011 04:08:38.721153 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-sync-sdg2s" event={"ID":"1d249b36-580e-424d-8f36-e366c932411c","Type":"ContainerDied","Data":"02bf66457fc098fe15c42a14af57eb113a56800408703ed35ed3a2b4f07ab74e"} Oct 11 04:08:39 crc kubenswrapper[4967]: I1011 04:08:39.137226 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6b6c6c8584-4qwkb" podUID="9c6c522c-bd96-42f6-b0ae-27d55d3121c4" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.156:9311/healthcheck\": read tcp 10.217.0.2:40402->10.217.0.156:9311: read: connection reset by peer" Oct 11 04:08:39 crc kubenswrapper[4967]: I1011 04:08:39.137253 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6b6c6c8584-4qwkb" podUID="9c6c522c-bd96-42f6-b0ae-27d55d3121c4" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.156:9311/healthcheck\": read tcp 10.217.0.2:40392->10.217.0.156:9311: read: connection reset by peer" Oct 11 04:08:39 crc kubenswrapper[4967]: I1011 04:08:39.744633 4967 generic.go:334] "Generic (PLEG): container finished" podID="9c6c522c-bd96-42f6-b0ae-27d55d3121c4" containerID="fab128d00e2d02acf1adf0f691a5c9d7a4302c62e8039f94ba4a887ed5d74b69" exitCode=0 Oct 11 04:08:39 crc kubenswrapper[4967]: I1011 04:08:39.745297 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b6c6c8584-4qwkb" event={"ID":"9c6c522c-bd96-42f6-b0ae-27d55d3121c4","Type":"ContainerDied","Data":"fab128d00e2d02acf1adf0f691a5c9d7a4302c62e8039f94ba4a887ed5d74b69"} Oct 11 04:08:41 crc kubenswrapper[4967]: I1011 04:08:41.300399 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6486584599-dfqxj" Oct 11 04:08:41 crc kubenswrapper[4967]: I1011 04:08:41.303101 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/swift-proxy-6486584599-dfqxj" Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.043031 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6b6c6c8584-4qwkb" podUID="9c6c522c-bd96-42f6-b0ae-27d55d3121c4" containerName="barbican-api-log" probeResult="failure" output="Get \"http://10.217.0.156:9311/healthcheck\": dial tcp 10.217.0.156:9311: connect: connection refused" Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.043040 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/barbican-api-6b6c6c8584-4qwkb" podUID="9c6c522c-bd96-42f6-b0ae-27d55d3121c4" containerName="barbican-api" probeResult="failure" output="Get \"http://10.217.0.156:9311/healthcheck\": dial tcp 10.217.0.156:9311: connect: connection refused" Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.517814 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-sync-sdg2s" Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.616206 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6b6c6c8584-4qwkb" Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.639923 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d249b36-580e-424d-8f36-e366c932411c-combined-ca-bundle\") pod \"1d249b36-580e-424d-8f36-e366c932411c\" (UID: \"1d249b36-580e-424d-8f36-e366c932411c\") " Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.640024 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-75mhg\" (UniqueName: \"kubernetes.io/projected/1d249b36-580e-424d-8f36-e366c932411c-kube-api-access-75mhg\") pod \"1d249b36-580e-424d-8f36-e366c932411c\" (UID: \"1d249b36-580e-424d-8f36-e366c932411c\") " Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.640062 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/1d249b36-580e-424d-8f36-e366c932411c-config-data-merged\") pod \"1d249b36-580e-424d-8f36-e366c932411c\" (UID: \"1d249b36-580e-424d-8f36-e366c932411c\") " Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.640154 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/1d249b36-580e-424d-8f36-e366c932411c-etc-podinfo\") pod \"1d249b36-580e-424d-8f36-e366c932411c\" (UID: \"1d249b36-580e-424d-8f36-e366c932411c\") " Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.640193 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d249b36-580e-424d-8f36-e366c932411c-scripts\") pod \"1d249b36-580e-424d-8f36-e366c932411c\" (UID: \"1d249b36-580e-424d-8f36-e366c932411c\") " Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.640255 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d249b36-580e-424d-8f36-e366c932411c-config-data\") pod \"1d249b36-580e-424d-8f36-e366c932411c\" (UID: \"1d249b36-580e-424d-8f36-e366c932411c\") " Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.640876 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d249b36-580e-424d-8f36-e366c932411c-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "1d249b36-580e-424d-8f36-e366c932411c" (UID: "1d249b36-580e-424d-8f36-e366c932411c"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.648424 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d249b36-580e-424d-8f36-e366c932411c-scripts" (OuterVolumeSpecName: "scripts") pod "1d249b36-580e-424d-8f36-e366c932411c" (UID: "1d249b36-580e-424d-8f36-e366c932411c"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.657678 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/1d249b36-580e-424d-8f36-e366c932411c-etc-podinfo" (OuterVolumeSpecName: "etc-podinfo") pod "1d249b36-580e-424d-8f36-e366c932411c" (UID: "1d249b36-580e-424d-8f36-e366c932411c"). InnerVolumeSpecName "etc-podinfo". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.657879 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d249b36-580e-424d-8f36-e366c932411c-kube-api-access-75mhg" (OuterVolumeSpecName: "kube-api-access-75mhg") pod "1d249b36-580e-424d-8f36-e366c932411c" (UID: "1d249b36-580e-424d-8f36-e366c932411c"). InnerVolumeSpecName "kube-api-access-75mhg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.679392 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d249b36-580e-424d-8f36-e366c932411c-config-data" (OuterVolumeSpecName: "config-data") pod "1d249b36-580e-424d-8f36-e366c932411c" (UID: "1d249b36-580e-424d-8f36-e366c932411c"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.717195 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d249b36-580e-424d-8f36-e366c932411c-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1d249b36-580e-424d-8f36-e366c932411c" (UID: "1d249b36-580e-424d-8f36-e366c932411c"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.741872 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c6c522c-bd96-42f6-b0ae-27d55d3121c4-combined-ca-bundle\") pod \"9c6c522c-bd96-42f6-b0ae-27d55d3121c4\" (UID: \"9c6c522c-bd96-42f6-b0ae-27d55d3121c4\") " Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.741943 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9c6c522c-bd96-42f6-b0ae-27d55d3121c4-config-data-custom\") pod \"9c6c522c-bd96-42f6-b0ae-27d55d3121c4\" (UID: \"9c6c522c-bd96-42f6-b0ae-27d55d3121c4\") " Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.742337 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h95fg\" (UniqueName: \"kubernetes.io/projected/9c6c522c-bd96-42f6-b0ae-27d55d3121c4-kube-api-access-h95fg\") pod \"9c6c522c-bd96-42f6-b0ae-27d55d3121c4\" (UID: \"9c6c522c-bd96-42f6-b0ae-27d55d3121c4\") " Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.742377 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c6c522c-bd96-42f6-b0ae-27d55d3121c4-logs\") pod \"9c6c522c-bd96-42f6-b0ae-27d55d3121c4\" (UID: \"9c6c522c-bd96-42f6-b0ae-27d55d3121c4\") " Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.742461 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c6c522c-bd96-42f6-b0ae-27d55d3121c4-config-data\") pod \"9c6c522c-bd96-42f6-b0ae-27d55d3121c4\" (UID: \"9c6c522c-bd96-42f6-b0ae-27d55d3121c4\") " Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.742979 4967 reconciler_common.go:293] "Volume detached for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/1d249b36-580e-424d-8f36-e366c932411c-etc-podinfo\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.742999 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d249b36-580e-424d-8f36-e366c932411c-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.743010 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1d249b36-580e-424d-8f36-e366c932411c-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.743022 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d249b36-580e-424d-8f36-e366c932411c-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.743033 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-75mhg\" (UniqueName: \"kubernetes.io/projected/1d249b36-580e-424d-8f36-e366c932411c-kube-api-access-75mhg\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.743047 4967 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/1d249b36-580e-424d-8f36-e366c932411c-config-data-merged\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.743232 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/9c6c522c-bd96-42f6-b0ae-27d55d3121c4-logs" (OuterVolumeSpecName: "logs") pod "9c6c522c-bd96-42f6-b0ae-27d55d3121c4" (UID: "9c6c522c-bd96-42f6-b0ae-27d55d3121c4"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.744878 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c6c522c-bd96-42f6-b0ae-27d55d3121c4-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "9c6c522c-bd96-42f6-b0ae-27d55d3121c4" (UID: "9c6c522c-bd96-42f6-b0ae-27d55d3121c4"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.747602 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9c6c522c-bd96-42f6-b0ae-27d55d3121c4-kube-api-access-h95fg" (OuterVolumeSpecName: "kube-api-access-h95fg") pod "9c6c522c-bd96-42f6-b0ae-27d55d3121c4" (UID: "9c6c522c-bd96-42f6-b0ae-27d55d3121c4"). InnerVolumeSpecName "kube-api-access-h95fg". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.773230 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c6c522c-bd96-42f6-b0ae-27d55d3121c4-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9c6c522c-bd96-42f6-b0ae-27d55d3121c4" (UID: "9c6c522c-bd96-42f6-b0ae-27d55d3121c4"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.788360 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/openstackclient" event={"ID":"23c5100e-07d0-4897-98de-f4e91329085e","Type":"ContainerStarted","Data":"ef5fbed065132c235456bf4c3ef48c24bfc01e8aff3cf047a98aca321b21c337"} Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.794079 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/barbican-api-6b6c6c8584-4qwkb" event={"ID":"9c6c522c-bd96-42f6-b0ae-27d55d3121c4","Type":"ContainerDied","Data":"782fd2c4860269e8a59f9b8daf4a6230aa99ba0245156609df6fa27e88f34205"} Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.794131 4967 scope.go:117] "RemoveContainer" containerID="fab128d00e2d02acf1adf0f691a5c9d7a4302c62e8039f94ba4a887ed5d74b69" Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.794147 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/barbican-api-6b6c6c8584-4qwkb" Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.798368 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8e8fb63d-44cd-427e-8324-743d83fc8416","Type":"ContainerStarted","Data":"f21cdca371f1bb647b12cf6cb96e7d808e5322b3a6d444b6285a714cf3b2f7b0"} Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.805607 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9c6c522c-bd96-42f6-b0ae-27d55d3121c4-config-data" (OuterVolumeSpecName: "config-data") pod "9c6c522c-bd96-42f6-b0ae-27d55d3121c4" (UID: "9c6c522c-bd96-42f6-b0ae-27d55d3121c4"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.808513 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/openstackclient" podStartSLOduration=3.028830565 podStartE2EDuration="17.808496861s" podCreationTimestamp="2025-10-11 04:08:25 +0000 UTC" firstStartedPulling="2025-10-11 04:08:27.571622365 +0000 UTC m=+1035.534831338" lastFinishedPulling="2025-10-11 04:08:42.351288701 +0000 UTC m=+1050.314497634" observedRunningTime="2025-10-11 04:08:42.806664829 +0000 UTC m=+1050.769873762" watchObservedRunningTime="2025-10-11 04:08:42.808496861 +0000 UTC m=+1050.771705794" Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.813762 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-db-sync-sdg2s" event={"ID":"1d249b36-580e-424d-8f36-e366c932411c","Type":"ContainerDied","Data":"76a842317569015afefcd4c722bd5d59e42b98c49beaa46907bf45a29d5ca90d"} Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.813806 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="76a842317569015afefcd4c722bd5d59e42b98c49beaa46907bf45a29d5ca90d" Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.813869 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-db-sync-sdg2s" Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.836192 4967 scope.go:117] "RemoveContainer" containerID="2713f1c59d64482f421e2cd9da89a876d483393a3a347f05ed054178dad43dfa" Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.847857 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9c6c522c-bd96-42f6-b0ae-27d55d3121c4-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.851617 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9c6c522c-bd96-42f6-b0ae-27d55d3121c4-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.851818 4967 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/9c6c522c-bd96-42f6-b0ae-27d55d3121c4-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.851880 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h95fg\" (UniqueName: \"kubernetes.io/projected/9c6c522c-bd96-42f6-b0ae-27d55d3121c4-kube-api-access-h95fg\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:42 crc kubenswrapper[4967]: I1011 04:08:42.851978 4967 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/9c6c522c-bd96-42f6-b0ae-27d55d3121c4-logs\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:43 crc kubenswrapper[4967]: I1011 04:08:43.118948 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-api-6b6c6c8584-4qwkb"] Oct 11 04:08:43 crc kubenswrapper[4967]: I1011 04:08:43.125756 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-api-6b6c6c8584-4qwkb"] Oct 11 04:08:43 crc kubenswrapper[4967]: I1011 04:08:43.825488 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6633fd1a-cd77-429d-b24f-33cb1bcba3f3","Type":"ContainerStarted","Data":"f4d63601c949bfd89d5966d88b6b8efa585c95b72f68c273a5331e5bc4404c23"} Oct 11 04:08:43 crc kubenswrapper[4967]: I1011 04:08:43.825850 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="6633fd1a-cd77-429d-b24f-33cb1bcba3f3" containerName="cinder-api-log" containerID="cri-o://0e9d2174e592c5fa8e61611c5b8336748ff5343d122e6c90769747d418b2208d" gracePeriod=30 Oct 11 04:08:43 crc kubenswrapper[4967]: I1011 04:08:43.825926 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 11 04:08:43 crc kubenswrapper[4967]: I1011 04:08:43.825968 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-api-0" podUID="6633fd1a-cd77-429d-b24f-33cb1bcba3f3" containerName="cinder-api" containerID="cri-o://f4d63601c949bfd89d5966d88b6b8efa585c95b72f68c273a5331e5bc4404c23" gracePeriod=30 Oct 11 04:08:43 crc kubenswrapper[4967]: I1011 04:08:43.838222 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8e8fb63d-44cd-427e-8324-743d83fc8416","Type":"ContainerStarted","Data":"013addb6d98525fab2a9a93cb773a0dd9d6766722ae77ed6790a40a2f9919f9e"} Oct 11 04:08:43 crc kubenswrapper[4967]: I1011 04:08:43.841622 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4bdd15be-7524-4447-bab5-43c0a6b40cf0","Type":"ContainerStarted","Data":"3c1ad1c69eef5590d20f4be3aaf48cd73f8aff961e369759999c3c0b227a7026"} Oct 11 04:08:43 crc kubenswrapper[4967]: I1011 04:08:43.861637 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=10.861616931 podStartE2EDuration="10.861616931s" podCreationTimestamp="2025-10-11 04:08:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:08:43.854707728 +0000 UTC m=+1051.817916661" watchObservedRunningTime="2025-10-11 04:08:43.861616931 +0000 UTC m=+1051.824825864" Oct 11 04:08:43 crc kubenswrapper[4967]: I1011 04:08:43.888477 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=9.982237593 podStartE2EDuration="10.888459997s" podCreationTimestamp="2025-10-11 04:08:33 +0000 UTC" firstStartedPulling="2025-10-11 04:08:34.673034898 +0000 UTC m=+1042.636243831" lastFinishedPulling="2025-10-11 04:08:35.579257302 +0000 UTC m=+1043.542466235" observedRunningTime="2025-10-11 04:08:43.884200941 +0000 UTC m=+1051.847409874" watchObservedRunningTime="2025-10-11 04:08:43.888459997 +0000 UTC m=+1051.851668930" Oct 11 04:08:43 crc kubenswrapper[4967]: I1011 04:08:43.948414 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-inspector-db-create-bprvt"] Oct 11 04:08:43 crc kubenswrapper[4967]: E1011 04:08:43.949264 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d249b36-580e-424d-8f36-e366c932411c" containerName="init" Oct 11 04:08:43 crc kubenswrapper[4967]: I1011 04:08:43.949366 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d249b36-580e-424d-8f36-e366c932411c" containerName="init" Oct 11 04:08:43 crc kubenswrapper[4967]: E1011 04:08:43.949425 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c6c522c-bd96-42f6-b0ae-27d55d3121c4" containerName="barbican-api-log" Oct 11 04:08:43 crc kubenswrapper[4967]: I1011 04:08:43.949480 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c6c522c-bd96-42f6-b0ae-27d55d3121c4" containerName="barbican-api-log" Oct 11 04:08:43 crc kubenswrapper[4967]: E1011 04:08:43.949539 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9c6c522c-bd96-42f6-b0ae-27d55d3121c4" containerName="barbican-api" Oct 11 04:08:43 crc kubenswrapper[4967]: I1011 04:08:43.949594 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="9c6c522c-bd96-42f6-b0ae-27d55d3121c4" containerName="barbican-api" Oct 11 04:08:43 crc kubenswrapper[4967]: E1011 04:08:43.949654 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d249b36-580e-424d-8f36-e366c932411c" containerName="ironic-db-sync" Oct 11 04:08:43 crc kubenswrapper[4967]: I1011 04:08:43.949702 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d249b36-580e-424d-8f36-e366c932411c" containerName="ironic-db-sync" Oct 11 04:08:43 crc kubenswrapper[4967]: I1011 04:08:43.949924 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d249b36-580e-424d-8f36-e366c932411c" containerName="ironic-db-sync" Oct 11 04:08:43 crc kubenswrapper[4967]: I1011 04:08:43.949994 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c6c522c-bd96-42f6-b0ae-27d55d3121c4" containerName="barbican-api-log" Oct 11 04:08:43 crc kubenswrapper[4967]: I1011 04:08:43.950052 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="9c6c522c-bd96-42f6-b0ae-27d55d3121c4" containerName="barbican-api" Oct 11 04:08:43 crc kubenswrapper[4967]: I1011 04:08:43.950649 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-create-bprvt" Oct 11 04:08:43 crc kubenswrapper[4967]: I1011 04:08:43.985724 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-db-create-bprvt"] Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.004154 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-neutron-agent-657f6548d9-9fb8w"] Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.018392 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.027183 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-ironic-neutron-agent-config-data" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.027394 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-ironic-dockercfg-vjmsm" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.077286 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4w9w8\" (UniqueName: \"kubernetes.io/projected/4390a950-7185-4fc0-b152-b5927940032b-kube-api-access-4w9w8\") pod \"ironic-neutron-agent-657f6548d9-9fb8w\" (UID: \"4390a950-7185-4fc0-b152-b5927940032b\") " pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.077399 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4390a950-7185-4fc0-b152-b5927940032b-combined-ca-bundle\") pod \"ironic-neutron-agent-657f6548d9-9fb8w\" (UID: \"4390a950-7185-4fc0-b152-b5927940032b\") " pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.077441 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4390a950-7185-4fc0-b152-b5927940032b-config\") pod \"ironic-neutron-agent-657f6548d9-9fb8w\" (UID: \"4390a950-7185-4fc0-b152-b5927940032b\") " pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.077779 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dqcl7\" (UniqueName: \"kubernetes.io/projected/12c05ddc-4b39-4924-924c-955b8639a839-kube-api-access-dqcl7\") pod \"ironic-inspector-db-create-bprvt\" (UID: \"12c05ddc-4b39-4924-924c-955b8639a839\") " pod="openstack/ironic-inspector-db-create-bprvt" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.091261 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-neutron-agent-657f6548d9-9fb8w"] Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.100380 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.098806 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/cinder-scheduler-0" podUID="4bdd15be-7524-4447-bab5-43c0a6b40cf0" containerName="cinder-scheduler" probeResult="failure" output="Get \"http://10.217.0.164:8080/\": dial tcp 10.217.0.164:8080: connect: connection refused" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.132422 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-5b5f75f756-jlhwg"] Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.134413 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-5b5f75f756-jlhwg" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.144945 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-api-config-data" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.149135 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-api-scripts" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.153441 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-config-data" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.159051 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-5b5f75f756-jlhwg"] Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.177469 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-6578955fd5-vhtnn" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.178932 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4w9w8\" (UniqueName: \"kubernetes.io/projected/4390a950-7185-4fc0-b152-b5927940032b-kube-api-access-4w9w8\") pod \"ironic-neutron-agent-657f6548d9-9fb8w\" (UID: \"4390a950-7185-4fc0-b152-b5927940032b\") " pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.178963 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c82b179-3510-4fb2-9352-d41e7193d221-logs\") pod \"ironic-5b5f75f756-jlhwg\" (UID: \"5c82b179-3510-4fb2-9352-d41e7193d221\") " pod="openstack/ironic-5b5f75f756-jlhwg" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.178990 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c82b179-3510-4fb2-9352-d41e7193d221-scripts\") pod \"ironic-5b5f75f756-jlhwg\" (UID: \"5c82b179-3510-4fb2-9352-d41e7193d221\") " pod="openstack/ironic-5b5f75f756-jlhwg" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.179032 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c82b179-3510-4fb2-9352-d41e7193d221-combined-ca-bundle\") pod \"ironic-5b5f75f756-jlhwg\" (UID: \"5c82b179-3510-4fb2-9352-d41e7193d221\") " pod="openstack/ironic-5b5f75f756-jlhwg" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.179061 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4390a950-7185-4fc0-b152-b5927940032b-combined-ca-bundle\") pod \"ironic-neutron-agent-657f6548d9-9fb8w\" (UID: \"4390a950-7185-4fc0-b152-b5927940032b\") " pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.179097 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/5c82b179-3510-4fb2-9352-d41e7193d221-config-data-merged\") pod \"ironic-5b5f75f756-jlhwg\" (UID: \"5c82b179-3510-4fb2-9352-d41e7193d221\") " pod="openstack/ironic-5b5f75f756-jlhwg" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.179121 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/4390a950-7185-4fc0-b152-b5927940032b-config\") pod \"ironic-neutron-agent-657f6548d9-9fb8w\" (UID: \"4390a950-7185-4fc0-b152-b5927940032b\") " pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.179149 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5c82b179-3510-4fb2-9352-d41e7193d221-config-data-custom\") pod \"ironic-5b5f75f756-jlhwg\" (UID: \"5c82b179-3510-4fb2-9352-d41e7193d221\") " pod="openstack/ironic-5b5f75f756-jlhwg" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.179190 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/5c82b179-3510-4fb2-9352-d41e7193d221-etc-podinfo\") pod \"ironic-5b5f75f756-jlhwg\" (UID: \"5c82b179-3510-4fb2-9352-d41e7193d221\") " pod="openstack/ironic-5b5f75f756-jlhwg" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.179208 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjn7n\" (UniqueName: \"kubernetes.io/projected/5c82b179-3510-4fb2-9352-d41e7193d221-kube-api-access-hjn7n\") pod \"ironic-5b5f75f756-jlhwg\" (UID: \"5c82b179-3510-4fb2-9352-d41e7193d221\") " pod="openstack/ironic-5b5f75f756-jlhwg" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.179224 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c82b179-3510-4fb2-9352-d41e7193d221-config-data\") pod \"ironic-5b5f75f756-jlhwg\" (UID: \"5c82b179-3510-4fb2-9352-d41e7193d221\") " pod="openstack/ironic-5b5f75f756-jlhwg" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.179257 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dqcl7\" (UniqueName: \"kubernetes.io/projected/12c05ddc-4b39-4924-924c-955b8639a839-kube-api-access-dqcl7\") pod \"ironic-inspector-db-create-bprvt\" (UID: \"12c05ddc-4b39-4924-924c-955b8639a839\") " pod="openstack/ironic-inspector-db-create-bprvt" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.184712 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/4390a950-7185-4fc0-b152-b5927940032b-config\") pod \"ironic-neutron-agent-657f6548d9-9fb8w\" (UID: \"4390a950-7185-4fc0-b152-b5927940032b\") " pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.185600 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4390a950-7185-4fc0-b152-b5927940032b-combined-ca-bundle\") pod \"ironic-neutron-agent-657f6548d9-9fb8w\" (UID: \"4390a950-7185-4fc0-b152-b5927940032b\") " pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.204546 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dqcl7\" (UniqueName: \"kubernetes.io/projected/12c05ddc-4b39-4924-924c-955b8639a839-kube-api-access-dqcl7\") pod \"ironic-inspector-db-create-bprvt\" (UID: \"12c05ddc-4b39-4924-924c-955b8639a839\") " pod="openstack/ironic-inspector-db-create-bprvt" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.292475 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-create-bprvt" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.293058 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4w9w8\" (UniqueName: \"kubernetes.io/projected/4390a950-7185-4fc0-b152-b5927940032b-kube-api-access-4w9w8\") pod \"ironic-neutron-agent-657f6548d9-9fb8w\" (UID: \"4390a950-7185-4fc0-b152-b5927940032b\") " pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.293295 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/5c82b179-3510-4fb2-9352-d41e7193d221-config-data-merged\") pod \"ironic-5b5f75f756-jlhwg\" (UID: \"5c82b179-3510-4fb2-9352-d41e7193d221\") " pod="openstack/ironic-5b5f75f756-jlhwg" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.293367 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5c82b179-3510-4fb2-9352-d41e7193d221-config-data-custom\") pod \"ironic-5b5f75f756-jlhwg\" (UID: \"5c82b179-3510-4fb2-9352-d41e7193d221\") " pod="openstack/ironic-5b5f75f756-jlhwg" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.293407 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/5c82b179-3510-4fb2-9352-d41e7193d221-etc-podinfo\") pod \"ironic-5b5f75f756-jlhwg\" (UID: \"5c82b179-3510-4fb2-9352-d41e7193d221\") " pod="openstack/ironic-5b5f75f756-jlhwg" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.293437 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjn7n\" (UniqueName: \"kubernetes.io/projected/5c82b179-3510-4fb2-9352-d41e7193d221-kube-api-access-hjn7n\") pod \"ironic-5b5f75f756-jlhwg\" (UID: \"5c82b179-3510-4fb2-9352-d41e7193d221\") " pod="openstack/ironic-5b5f75f756-jlhwg" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.293453 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c82b179-3510-4fb2-9352-d41e7193d221-config-data\") pod \"ironic-5b5f75f756-jlhwg\" (UID: \"5c82b179-3510-4fb2-9352-d41e7193d221\") " pod="openstack/ironic-5b5f75f756-jlhwg" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.293519 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c82b179-3510-4fb2-9352-d41e7193d221-logs\") pod \"ironic-5b5f75f756-jlhwg\" (UID: \"5c82b179-3510-4fb2-9352-d41e7193d221\") " pod="openstack/ironic-5b5f75f756-jlhwg" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.293547 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c82b179-3510-4fb2-9352-d41e7193d221-scripts\") pod \"ironic-5b5f75f756-jlhwg\" (UID: \"5c82b179-3510-4fb2-9352-d41e7193d221\") " pod="openstack/ironic-5b5f75f756-jlhwg" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.293607 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c82b179-3510-4fb2-9352-d41e7193d221-combined-ca-bundle\") pod \"ironic-5b5f75f756-jlhwg\" (UID: \"5c82b179-3510-4fb2-9352-d41e7193d221\") " pod="openstack/ironic-5b5f75f756-jlhwg" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.295527 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/5c82b179-3510-4fb2-9352-d41e7193d221-config-data-merged\") pod \"ironic-5b5f75f756-jlhwg\" (UID: \"5c82b179-3510-4fb2-9352-d41e7193d221\") " pod="openstack/ironic-5b5f75f756-jlhwg" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.300847 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c82b179-3510-4fb2-9352-d41e7193d221-scripts\") pod \"ironic-5b5f75f756-jlhwg\" (UID: \"5c82b179-3510-4fb2-9352-d41e7193d221\") " pod="openstack/ironic-5b5f75f756-jlhwg" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.301400 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5c82b179-3510-4fb2-9352-d41e7193d221-config-data-custom\") pod \"ironic-5b5f75f756-jlhwg\" (UID: \"5c82b179-3510-4fb2-9352-d41e7193d221\") " pod="openstack/ironic-5b5f75f756-jlhwg" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.305615 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/5c82b179-3510-4fb2-9352-d41e7193d221-etc-podinfo\") pod \"ironic-5b5f75f756-jlhwg\" (UID: \"5c82b179-3510-4fb2-9352-d41e7193d221\") " pod="openstack/ironic-5b5f75f756-jlhwg" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.320109 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c82b179-3510-4fb2-9352-d41e7193d221-logs\") pod \"ironic-5b5f75f756-jlhwg\" (UID: \"5c82b179-3510-4fb2-9352-d41e7193d221\") " pod="openstack/ironic-5b5f75f756-jlhwg" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.336746 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c82b179-3510-4fb2-9352-d41e7193d221-config-data\") pod \"ironic-5b5f75f756-jlhwg\" (UID: \"5c82b179-3510-4fb2-9352-d41e7193d221\") " pod="openstack/ironic-5b5f75f756-jlhwg" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.343610 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c82b179-3510-4fb2-9352-d41e7193d221-combined-ca-bundle\") pod \"ironic-5b5f75f756-jlhwg\" (UID: \"5c82b179-3510-4fb2-9352-d41e7193d221\") " pod="openstack/ironic-5b5f75f756-jlhwg" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.355730 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjn7n\" (UniqueName: \"kubernetes.io/projected/5c82b179-3510-4fb2-9352-d41e7193d221-kube-api-access-hjn7n\") pod \"ironic-5b5f75f756-jlhwg\" (UID: \"5c82b179-3510-4fb2-9352-d41e7193d221\") " pod="openstack/ironic-5b5f75f756-jlhwg" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.368278 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-jdrjp"] Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.368507 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-848cf88cfc-jdrjp" podUID="7a6e1897-5fd1-4c34-8f26-cd97ced1eb43" containerName="dnsmasq-dns" containerID="cri-o://21c9768d6219505f16dbf91b5cc918947d01146763a1e6c0d31c780ff517d5c9" gracePeriod=10 Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.441627 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.475583 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-5b5f75f756-jlhwg" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.831594 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9c6c522c-bd96-42f6-b0ae-27d55d3121c4" path="/var/lib/kubelet/pods/9c6c522c-bd96-42f6-b0ae-27d55d3121c4/volumes" Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.875041 4967 generic.go:334] "Generic (PLEG): container finished" podID="6633fd1a-cd77-429d-b24f-33cb1bcba3f3" containerID="f4d63601c949bfd89d5966d88b6b8efa585c95b72f68c273a5331e5bc4404c23" exitCode=0 Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.875111 4967 generic.go:334] "Generic (PLEG): container finished" podID="6633fd1a-cd77-429d-b24f-33cb1bcba3f3" containerID="0e9d2174e592c5fa8e61611c5b8336748ff5343d122e6c90769747d418b2208d" exitCode=143 Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.875168 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6633fd1a-cd77-429d-b24f-33cb1bcba3f3","Type":"ContainerDied","Data":"f4d63601c949bfd89d5966d88b6b8efa585c95b72f68c273a5331e5bc4404c23"} Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.875203 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6633fd1a-cd77-429d-b24f-33cb1bcba3f3","Type":"ContainerDied","Data":"0e9d2174e592c5fa8e61611c5b8336748ff5343d122e6c90769747d418b2208d"} Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.877287 4967 generic.go:334] "Generic (PLEG): container finished" podID="7a6e1897-5fd1-4c34-8f26-cd97ced1eb43" containerID="21c9768d6219505f16dbf91b5cc918947d01146763a1e6c0d31c780ff517d5c9" exitCode=0 Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.877385 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-jdrjp" event={"ID":"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43","Type":"ContainerDied","Data":"21c9768d6219505f16dbf91b5cc918947d01146763a1e6c0d31c780ff517d5c9"} Oct 11 04:08:44 crc kubenswrapper[4967]: I1011 04:08:44.880777 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8e8fb63d-44cd-427e-8324-743d83fc8416","Type":"ContainerStarted","Data":"3021661bce4ef00784088c3e9a64d812c8301954bd3c35ac594364a201d895ee"} Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.071732 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-db-create-bprvt"] Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.087406 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-conductor-0"] Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.095815 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-conductor-0" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.106191 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-conductor-scripts" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.106417 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-conductor-config-data" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.127438 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-conductor-0"] Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.201795 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-jdrjp" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.209393 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-neutron-agent-657f6548d9-9fb8w"] Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.214605 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.225437 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/64c75ea6-aed6-4bf6-8eef-39e76384e14f-config-data-custom\") pod \"ironic-conductor-0\" (UID: \"64c75ea6-aed6-4bf6-8eef-39e76384e14f\") " pod="openstack/ironic-conductor-0" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.225522 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/64c75ea6-aed6-4bf6-8eef-39e76384e14f-etc-podinfo\") pod \"ironic-conductor-0\" (UID: \"64c75ea6-aed6-4bf6-8eef-39e76384e14f\") " pod="openstack/ironic-conductor-0" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.225548 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qn74m\" (UniqueName: \"kubernetes.io/projected/64c75ea6-aed6-4bf6-8eef-39e76384e14f-kube-api-access-qn74m\") pod \"ironic-conductor-0\" (UID: \"64c75ea6-aed6-4bf6-8eef-39e76384e14f\") " pod="openstack/ironic-conductor-0" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.225663 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ironic-conductor-0\" (UID: \"64c75ea6-aed6-4bf6-8eef-39e76384e14f\") " pod="openstack/ironic-conductor-0" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.225767 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64c75ea6-aed6-4bf6-8eef-39e76384e14f-combined-ca-bundle\") pod \"ironic-conductor-0\" (UID: \"64c75ea6-aed6-4bf6-8eef-39e76384e14f\") " pod="openstack/ironic-conductor-0" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.225822 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64c75ea6-aed6-4bf6-8eef-39e76384e14f-config-data\") pod \"ironic-conductor-0\" (UID: \"64c75ea6-aed6-4bf6-8eef-39e76384e14f\") " pod="openstack/ironic-conductor-0" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.225844 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/64c75ea6-aed6-4bf6-8eef-39e76384e14f-scripts\") pod \"ironic-conductor-0\" (UID: \"64c75ea6-aed6-4bf6-8eef-39e76384e14f\") " pod="openstack/ironic-conductor-0" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.225918 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/64c75ea6-aed6-4bf6-8eef-39e76384e14f-config-data-merged\") pod \"ironic-conductor-0\" (UID: \"64c75ea6-aed6-4bf6-8eef-39e76384e14f\") " pod="openstack/ironic-conductor-0" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.327259 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-logs\") pod \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\" (UID: \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\") " Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.327316 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-config-data-custom\") pod \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\" (UID: \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\") " Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.327342 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-etc-machine-id\") pod \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\" (UID: \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\") " Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.327780 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-combined-ca-bundle\") pod \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\" (UID: \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\") " Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.327807 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-config-data\") pod \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\" (UID: \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\") " Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.327823 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hrx7l\" (UniqueName: \"kubernetes.io/projected/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-kube-api-access-hrx7l\") pod \"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43\" (UID: \"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43\") " Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.327850 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-scripts\") pod \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\" (UID: \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\") " Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.327959 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-ovsdbserver-sb\") pod \"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43\" (UID: \"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43\") " Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.327991 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-config\") pod \"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43\" (UID: \"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43\") " Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.328090 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-ovsdbserver-nb\") pod \"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43\" (UID: \"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43\") " Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.328116 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lr5wf\" (UniqueName: \"kubernetes.io/projected/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-kube-api-access-lr5wf\") pod \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\" (UID: \"6633fd1a-cd77-429d-b24f-33cb1bcba3f3\") " Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.328149 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-dns-svc\") pod \"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43\" (UID: \"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43\") " Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.328163 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-dns-swift-storage-0\") pod \"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43\" (UID: \"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43\") " Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.328352 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/64c75ea6-aed6-4bf6-8eef-39e76384e14f-etc-podinfo\") pod \"ironic-conductor-0\" (UID: \"64c75ea6-aed6-4bf6-8eef-39e76384e14f\") " pod="openstack/ironic-conductor-0" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.328369 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qn74m\" (UniqueName: \"kubernetes.io/projected/64c75ea6-aed6-4bf6-8eef-39e76384e14f-kube-api-access-qn74m\") pod \"ironic-conductor-0\" (UID: \"64c75ea6-aed6-4bf6-8eef-39e76384e14f\") " pod="openstack/ironic-conductor-0" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.328415 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ironic-conductor-0\" (UID: \"64c75ea6-aed6-4bf6-8eef-39e76384e14f\") " pod="openstack/ironic-conductor-0" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.328478 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64c75ea6-aed6-4bf6-8eef-39e76384e14f-combined-ca-bundle\") pod \"ironic-conductor-0\" (UID: \"64c75ea6-aed6-4bf6-8eef-39e76384e14f\") " pod="openstack/ironic-conductor-0" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.328510 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64c75ea6-aed6-4bf6-8eef-39e76384e14f-config-data\") pod \"ironic-conductor-0\" (UID: \"64c75ea6-aed6-4bf6-8eef-39e76384e14f\") " pod="openstack/ironic-conductor-0" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.328523 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/64c75ea6-aed6-4bf6-8eef-39e76384e14f-scripts\") pod \"ironic-conductor-0\" (UID: \"64c75ea6-aed6-4bf6-8eef-39e76384e14f\") " pod="openstack/ironic-conductor-0" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.328566 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/64c75ea6-aed6-4bf6-8eef-39e76384e14f-config-data-merged\") pod \"ironic-conductor-0\" (UID: \"64c75ea6-aed6-4bf6-8eef-39e76384e14f\") " pod="openstack/ironic-conductor-0" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.328592 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/64c75ea6-aed6-4bf6-8eef-39e76384e14f-config-data-custom\") pod \"ironic-conductor-0\" (UID: \"64c75ea6-aed6-4bf6-8eef-39e76384e14f\") " pod="openstack/ironic-conductor-0" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.332463 4967 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ironic-conductor-0\" (UID: \"64c75ea6-aed6-4bf6-8eef-39e76384e14f\") device mount path \"/mnt/openstack/pv05\"" pod="openstack/ironic-conductor-0" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.338550 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/64c75ea6-aed6-4bf6-8eef-39e76384e14f-scripts\") pod \"ironic-conductor-0\" (UID: \"64c75ea6-aed6-4bf6-8eef-39e76384e14f\") " pod="openstack/ironic-conductor-0" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.339660 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64c75ea6-aed6-4bf6-8eef-39e76384e14f-config-data\") pod \"ironic-conductor-0\" (UID: \"64c75ea6-aed6-4bf6-8eef-39e76384e14f\") " pod="openstack/ironic-conductor-0" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.339877 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/64c75ea6-aed6-4bf6-8eef-39e76384e14f-config-data-merged\") pod \"ironic-conductor-0\" (UID: \"64c75ea6-aed6-4bf6-8eef-39e76384e14f\") " pod="openstack/ironic-conductor-0" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.340198 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-logs" (OuterVolumeSpecName: "logs") pod "6633fd1a-cd77-429d-b24f-33cb1bcba3f3" (UID: "6633fd1a-cd77-429d-b24f-33cb1bcba3f3"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.340615 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/64c75ea6-aed6-4bf6-8eef-39e76384e14f-config-data-custom\") pod \"ironic-conductor-0\" (UID: \"64c75ea6-aed6-4bf6-8eef-39e76384e14f\") " pod="openstack/ironic-conductor-0" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.340706 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "6633fd1a-cd77-429d-b24f-33cb1bcba3f3" (UID: "6633fd1a-cd77-429d-b24f-33cb1bcba3f3"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.342164 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "6633fd1a-cd77-429d-b24f-33cb1bcba3f3" (UID: "6633fd1a-cd77-429d-b24f-33cb1bcba3f3"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.342530 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/64c75ea6-aed6-4bf6-8eef-39e76384e14f-etc-podinfo\") pod \"ironic-conductor-0\" (UID: \"64c75ea6-aed6-4bf6-8eef-39e76384e14f\") " pod="openstack/ironic-conductor-0" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.346902 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64c75ea6-aed6-4bf6-8eef-39e76384e14f-combined-ca-bundle\") pod \"ironic-conductor-0\" (UID: \"64c75ea6-aed6-4bf6-8eef-39e76384e14f\") " pod="openstack/ironic-conductor-0" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.351106 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qn74m\" (UniqueName: \"kubernetes.io/projected/64c75ea6-aed6-4bf6-8eef-39e76384e14f-kube-api-access-qn74m\") pod \"ironic-conductor-0\" (UID: \"64c75ea6-aed6-4bf6-8eef-39e76384e14f\") " pod="openstack/ironic-conductor-0" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.354824 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-kube-api-access-hrx7l" (OuterVolumeSpecName: "kube-api-access-hrx7l") pod "7a6e1897-5fd1-4c34-8f26-cd97ced1eb43" (UID: "7a6e1897-5fd1-4c34-8f26-cd97ced1eb43"). InnerVolumeSpecName "kube-api-access-hrx7l". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.354978 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-scripts" (OuterVolumeSpecName: "scripts") pod "6633fd1a-cd77-429d-b24f-33cb1bcba3f3" (UID: "6633fd1a-cd77-429d-b24f-33cb1bcba3f3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.364646 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-kube-api-access-lr5wf" (OuterVolumeSpecName: "kube-api-access-lr5wf") pod "6633fd1a-cd77-429d-b24f-33cb1bcba3f3" (UID: "6633fd1a-cd77-429d-b24f-33cb1bcba3f3"). InnerVolumeSpecName "kube-api-access-lr5wf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.374748 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage05-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage05-crc\") pod \"ironic-conductor-0\" (UID: \"64c75ea6-aed6-4bf6-8eef-39e76384e14f\") " pod="openstack/ironic-conductor-0" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.405389 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "7a6e1897-5fd1-4c34-8f26-cd97ced1eb43" (UID: "7a6e1897-5fd1-4c34-8f26-cd97ced1eb43"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.409191 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "6633fd1a-cd77-429d-b24f-33cb1bcba3f3" (UID: "6633fd1a-cd77-429d-b24f-33cb1bcba3f3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.411746 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-5b5f75f756-jlhwg"] Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.428470 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-conductor-0" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.430092 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.430201 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hrx7l\" (UniqueName: \"kubernetes.io/projected/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-kube-api-access-hrx7l\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.430266 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.430320 4967 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.430376 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lr5wf\" (UniqueName: \"kubernetes.io/projected/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-kube-api-access-lr5wf\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.430430 4967 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-logs\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.430491 4967 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.430607 4967 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.439267 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "7a6e1897-5fd1-4c34-8f26-cd97ced1eb43" (UID: "7a6e1897-5fd1-4c34-8f26-cd97ced1eb43"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.439942 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "7a6e1897-5fd1-4c34-8f26-cd97ced1eb43" (UID: "7a6e1897-5fd1-4c34-8f26-cd97ced1eb43"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.455169 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "7a6e1897-5fd1-4c34-8f26-cd97ced1eb43" (UID: "7a6e1897-5fd1-4c34-8f26-cd97ced1eb43"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.469423 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-config-data" (OuterVolumeSpecName: "config-data") pod "6633fd1a-cd77-429d-b24f-33cb1bcba3f3" (UID: "6633fd1a-cd77-429d-b24f-33cb1bcba3f3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.505851 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-config" (OuterVolumeSpecName: "config") pod "7a6e1897-5fd1-4c34-8f26-cd97ced1eb43" (UID: "7a6e1897-5fd1-4c34-8f26-cd97ced1eb43"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.532458 4967 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.532489 4967 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.532498 4967 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.532506 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/6633fd1a-cd77-429d-b24f-33cb1bcba3f3-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.532518 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43-config\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.892628 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" event={"ID":"4390a950-7185-4fc0-b152-b5927940032b","Type":"ContainerStarted","Data":"95faa5bca3092caf2639c6e0c172cc008bf7d679ff1cf315c648aa94997b403f"} Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.894203 4967 generic.go:334] "Generic (PLEG): container finished" podID="12c05ddc-4b39-4924-924c-955b8639a839" containerID="de98430f092cd5ed87090b453397b430785e49854d3b7844ef1aa6d36fcf2da2" exitCode=0 Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.894296 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-create-bprvt" event={"ID":"12c05ddc-4b39-4924-924c-955b8639a839","Type":"ContainerDied","Data":"de98430f092cd5ed87090b453397b430785e49854d3b7844ef1aa6d36fcf2da2"} Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.894348 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-create-bprvt" event={"ID":"12c05ddc-4b39-4924-924c-955b8639a839","Type":"ContainerStarted","Data":"df3d198f9ce2a0e29b4ac644d8045c0310c19106ada7e2dc3de68965b39c7de5"} Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.896194 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"6633fd1a-cd77-429d-b24f-33cb1bcba3f3","Type":"ContainerDied","Data":"931caf0e691edc8d915cc559cf6c851833dbc13584ef8af86015734bad24ea73"} Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.896212 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.896232 4967 scope.go:117] "RemoveContainer" containerID="f4d63601c949bfd89d5966d88b6b8efa585c95b72f68c273a5331e5bc4404c23" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.899485 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-848cf88cfc-jdrjp" event={"ID":"7a6e1897-5fd1-4c34-8f26-cd97ced1eb43","Type":"ContainerDied","Data":"40529b23e0dd5d3fd6c984724438049d3aebc7de61e95e71d71b12869b2e206b"} Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.899593 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-848cf88cfc-jdrjp" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.907234 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8e8fb63d-44cd-427e-8324-743d83fc8416","Type":"ContainerStarted","Data":"5202ac518e2e2e7aaca1d0784961a9f966e120ded550fc4bac7e0bcbbf151e49"} Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.909896 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-5b5f75f756-jlhwg" event={"ID":"5c82b179-3510-4fb2-9352-d41e7193d221","Type":"ContainerStarted","Data":"c278653c52e37dbf15ebd8eb218ca2dea6ad7489d8831b811d37f5ffc16ec9b9"} Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.918003 4967 scope.go:117] "RemoveContainer" containerID="0e9d2174e592c5fa8e61611c5b8336748ff5343d122e6c90769747d418b2208d" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.933024 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-api-0"] Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.947671 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-api-0"] Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.957947 4967 scope.go:117] "RemoveContainer" containerID="21c9768d6219505f16dbf91b5cc918947d01146763a1e6c0d31c780ff517d5c9" Oct 11 04:08:45 crc kubenswrapper[4967]: I1011 04:08:45.974813 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-api-0"] Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.001402 4967 scope.go:117] "RemoveContainer" containerID="88cebd6157e8cdeb6552dc56736b5a329542c1553a121919af966eab72804d6d" Oct 11 04:08:46 crc kubenswrapper[4967]: E1011 04:08:46.004612 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a6e1897-5fd1-4c34-8f26-cd97ced1eb43" containerName="init" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.004638 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a6e1897-5fd1-4c34-8f26-cd97ced1eb43" containerName="init" Oct 11 04:08:46 crc kubenswrapper[4967]: E1011 04:08:46.004686 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6633fd1a-cd77-429d-b24f-33cb1bcba3f3" containerName="cinder-api-log" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.004693 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="6633fd1a-cd77-429d-b24f-33cb1bcba3f3" containerName="cinder-api-log" Oct 11 04:08:46 crc kubenswrapper[4967]: E1011 04:08:46.004717 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6633fd1a-cd77-429d-b24f-33cb1bcba3f3" containerName="cinder-api" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.004724 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="6633fd1a-cd77-429d-b24f-33cb1bcba3f3" containerName="cinder-api" Oct 11 04:08:46 crc kubenswrapper[4967]: E1011 04:08:46.004752 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7a6e1897-5fd1-4c34-8f26-cd97ced1eb43" containerName="dnsmasq-dns" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.004759 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="7a6e1897-5fd1-4c34-8f26-cd97ced1eb43" containerName="dnsmasq-dns" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.007047 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="6633fd1a-cd77-429d-b24f-33cb1bcba3f3" containerName="cinder-api" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.007095 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="7a6e1897-5fd1-4c34-8f26-cd97ced1eb43" containerName="dnsmasq-dns" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.007158 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="6633fd1a-cd77-429d-b24f-33cb1bcba3f3" containerName="cinder-api-log" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.024119 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-jdrjp"] Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.024158 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-848cf88cfc-jdrjp"] Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.024250 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.030769 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.032615 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-api-config-data" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.032820 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-public-svc" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.033161 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-cinder-internal-svc" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.038349 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-conductor-0"] Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.157595 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c900849f-54f8-44ed-b759-ae6fbbe441da-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c900849f-54f8-44ed-b759-ae6fbbe441da\") " pod="openstack/cinder-api-0" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.158096 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c900849f-54f8-44ed-b759-ae6fbbe441da-logs\") pod \"cinder-api-0\" (UID: \"c900849f-54f8-44ed-b759-ae6fbbe441da\") " pod="openstack/cinder-api-0" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.158193 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-j268g\" (UniqueName: \"kubernetes.io/projected/c900849f-54f8-44ed-b759-ae6fbbe441da-kube-api-access-j268g\") pod \"cinder-api-0\" (UID: \"c900849f-54f8-44ed-b759-ae6fbbe441da\") " pod="openstack/cinder-api-0" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.158327 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c900849f-54f8-44ed-b759-ae6fbbe441da-config-data-custom\") pod \"cinder-api-0\" (UID: \"c900849f-54f8-44ed-b759-ae6fbbe441da\") " pod="openstack/cinder-api-0" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.158497 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c900849f-54f8-44ed-b759-ae6fbbe441da-scripts\") pod \"cinder-api-0\" (UID: \"c900849f-54f8-44ed-b759-ae6fbbe441da\") " pod="openstack/cinder-api-0" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.158626 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c900849f-54f8-44ed-b759-ae6fbbe441da-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c900849f-54f8-44ed-b759-ae6fbbe441da\") " pod="openstack/cinder-api-0" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.158727 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c900849f-54f8-44ed-b759-ae6fbbe441da-public-tls-certs\") pod \"cinder-api-0\" (UID: \"c900849f-54f8-44ed-b759-ae6fbbe441da\") " pod="openstack/cinder-api-0" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.158839 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c900849f-54f8-44ed-b759-ae6fbbe441da-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"c900849f-54f8-44ed-b759-ae6fbbe441da\") " pod="openstack/cinder-api-0" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.158926 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c900849f-54f8-44ed-b759-ae6fbbe441da-config-data\") pod \"cinder-api-0\" (UID: \"c900849f-54f8-44ed-b759-ae6fbbe441da\") " pod="openstack/cinder-api-0" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.263893 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c900849f-54f8-44ed-b759-ae6fbbe441da-public-tls-certs\") pod \"cinder-api-0\" (UID: \"c900849f-54f8-44ed-b759-ae6fbbe441da\") " pod="openstack/cinder-api-0" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.263968 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c900849f-54f8-44ed-b759-ae6fbbe441da-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"c900849f-54f8-44ed-b759-ae6fbbe441da\") " pod="openstack/cinder-api-0" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.263989 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c900849f-54f8-44ed-b759-ae6fbbe441da-config-data\") pod \"cinder-api-0\" (UID: \"c900849f-54f8-44ed-b759-ae6fbbe441da\") " pod="openstack/cinder-api-0" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.264089 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c900849f-54f8-44ed-b759-ae6fbbe441da-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c900849f-54f8-44ed-b759-ae6fbbe441da\") " pod="openstack/cinder-api-0" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.264158 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c900849f-54f8-44ed-b759-ae6fbbe441da-logs\") pod \"cinder-api-0\" (UID: \"c900849f-54f8-44ed-b759-ae6fbbe441da\") " pod="openstack/cinder-api-0" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.264184 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-j268g\" (UniqueName: \"kubernetes.io/projected/c900849f-54f8-44ed-b759-ae6fbbe441da-kube-api-access-j268g\") pod \"cinder-api-0\" (UID: \"c900849f-54f8-44ed-b759-ae6fbbe441da\") " pod="openstack/cinder-api-0" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.264244 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c900849f-54f8-44ed-b759-ae6fbbe441da-config-data-custom\") pod \"cinder-api-0\" (UID: \"c900849f-54f8-44ed-b759-ae6fbbe441da\") " pod="openstack/cinder-api-0" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.264277 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c900849f-54f8-44ed-b759-ae6fbbe441da-scripts\") pod \"cinder-api-0\" (UID: \"c900849f-54f8-44ed-b759-ae6fbbe441da\") " pod="openstack/cinder-api-0" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.264341 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c900849f-54f8-44ed-b759-ae6fbbe441da-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c900849f-54f8-44ed-b759-ae6fbbe441da\") " pod="openstack/cinder-api-0" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.264405 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/c900849f-54f8-44ed-b759-ae6fbbe441da-etc-machine-id\") pod \"cinder-api-0\" (UID: \"c900849f-54f8-44ed-b759-ae6fbbe441da\") " pod="openstack/cinder-api-0" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.264875 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/c900849f-54f8-44ed-b759-ae6fbbe441da-logs\") pod \"cinder-api-0\" (UID: \"c900849f-54f8-44ed-b759-ae6fbbe441da\") " pod="openstack/cinder-api-0" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.274010 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c900849f-54f8-44ed-b759-ae6fbbe441da-scripts\") pod \"cinder-api-0\" (UID: \"c900849f-54f8-44ed-b759-ae6fbbe441da\") " pod="openstack/cinder-api-0" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.275308 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c900849f-54f8-44ed-b759-ae6fbbe441da-config-data\") pod \"cinder-api-0\" (UID: \"c900849f-54f8-44ed-b759-ae6fbbe441da\") " pod="openstack/cinder-api-0" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.275447 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/c900849f-54f8-44ed-b759-ae6fbbe441da-config-data-custom\") pod \"cinder-api-0\" (UID: \"c900849f-54f8-44ed-b759-ae6fbbe441da\") " pod="openstack/cinder-api-0" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.275528 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c900849f-54f8-44ed-b759-ae6fbbe441da-internal-tls-certs\") pod \"cinder-api-0\" (UID: \"c900849f-54f8-44ed-b759-ae6fbbe441da\") " pod="openstack/cinder-api-0" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.275528 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c900849f-54f8-44ed-b759-ae6fbbe441da-public-tls-certs\") pod \"cinder-api-0\" (UID: \"c900849f-54f8-44ed-b759-ae6fbbe441da\") " pod="openstack/cinder-api-0" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.277873 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c900849f-54f8-44ed-b759-ae6fbbe441da-combined-ca-bundle\") pod \"cinder-api-0\" (UID: \"c900849f-54f8-44ed-b759-ae6fbbe441da\") " pod="openstack/cinder-api-0" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.288375 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-j268g\" (UniqueName: \"kubernetes.io/projected/c900849f-54f8-44ed-b759-ae6fbbe441da-kube-api-access-j268g\") pod \"cinder-api-0\" (UID: \"c900849f-54f8-44ed-b759-ae6fbbe441da\") " pod="openstack/cinder-api-0" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.358780 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-api-0" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.773625 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-85b58c7f48-9dgpg"] Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.775592 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.777433 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ironic-internal-svc" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.777808 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ironic-public-svc" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.782015 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-85b58c7f48-9dgpg"] Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.828815 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6633fd1a-cd77-429d-b24f-33cb1bcba3f3" path="/var/lib/kubelet/pods/6633fd1a-cd77-429d-b24f-33cb1bcba3f3/volumes" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.830186 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7a6e1897-5fd1-4c34-8f26-cd97ced1eb43" path="/var/lib/kubelet/pods/7a6e1897-5fd1-4c34-8f26-cd97ced1eb43/volumes" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.847817 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-api-0"] Oct 11 04:08:46 crc kubenswrapper[4967]: W1011 04:08:46.848316 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc900849f_54f8_44ed_b759_ae6fbbe441da.slice/crio-fd9a049da6109dae06db5e2f65a9433744f9cb3444c7f91fb3248f5592a31017 WatchSource:0}: Error finding container fd9a049da6109dae06db5e2f65a9433744f9cb3444c7f91fb3248f5592a31017: Status 404 returned error can't find the container with id fd9a049da6109dae06db5e2f65a9433744f9cb3444c7f91fb3248f5592a31017 Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.872353 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8ec340a-d11c-4b3b-9d2c-691df43b8892-combined-ca-bundle\") pod \"ironic-85b58c7f48-9dgpg\" (UID: \"f8ec340a-d11c-4b3b-9d2c-691df43b8892\") " pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.872446 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8ec340a-d11c-4b3b-9d2c-691df43b8892-public-tls-certs\") pod \"ironic-85b58c7f48-9dgpg\" (UID: \"f8ec340a-d11c-4b3b-9d2c-691df43b8892\") " pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.872484 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8ec340a-d11c-4b3b-9d2c-691df43b8892-internal-tls-certs\") pod \"ironic-85b58c7f48-9dgpg\" (UID: \"f8ec340a-d11c-4b3b-9d2c-691df43b8892\") " pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.872545 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-285wq\" (UniqueName: \"kubernetes.io/projected/f8ec340a-d11c-4b3b-9d2c-691df43b8892-kube-api-access-285wq\") pod \"ironic-85b58c7f48-9dgpg\" (UID: \"f8ec340a-d11c-4b3b-9d2c-691df43b8892\") " pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.872650 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f8ec340a-d11c-4b3b-9d2c-691df43b8892-config-data-merged\") pod \"ironic-85b58c7f48-9dgpg\" (UID: \"f8ec340a-d11c-4b3b-9d2c-691df43b8892\") " pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.872688 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8ec340a-d11c-4b3b-9d2c-691df43b8892-config-data\") pod \"ironic-85b58c7f48-9dgpg\" (UID: \"f8ec340a-d11c-4b3b-9d2c-691df43b8892\") " pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.872710 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8ec340a-d11c-4b3b-9d2c-691df43b8892-scripts\") pod \"ironic-85b58c7f48-9dgpg\" (UID: \"f8ec340a-d11c-4b3b-9d2c-691df43b8892\") " pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.872782 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8ec340a-d11c-4b3b-9d2c-691df43b8892-config-data-custom\") pod \"ironic-85b58c7f48-9dgpg\" (UID: \"f8ec340a-d11c-4b3b-9d2c-691df43b8892\") " pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.872819 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8ec340a-d11c-4b3b-9d2c-691df43b8892-logs\") pod \"ironic-85b58c7f48-9dgpg\" (UID: \"f8ec340a-d11c-4b3b-9d2c-691df43b8892\") " pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.872862 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/f8ec340a-d11c-4b3b-9d2c-691df43b8892-etc-podinfo\") pod \"ironic-85b58c7f48-9dgpg\" (UID: \"f8ec340a-d11c-4b3b-9d2c-691df43b8892\") " pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.924814 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c900849f-54f8-44ed-b759-ae6fbbe441da","Type":"ContainerStarted","Data":"fd9a049da6109dae06db5e2f65a9433744f9cb3444c7f91fb3248f5592a31017"} Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.931272 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"64c75ea6-aed6-4bf6-8eef-39e76384e14f","Type":"ContainerStarted","Data":"62a3e33ddd46e7b80643f3a4c852455565dfc189794219113986f6815eb3264b"} Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.931310 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"64c75ea6-aed6-4bf6-8eef-39e76384e14f","Type":"ContainerStarted","Data":"1d548604257adb1d6f324bb3348187150c55a8f6ec76582c584ddbb1dc241475"} Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.974153 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-285wq\" (UniqueName: \"kubernetes.io/projected/f8ec340a-d11c-4b3b-9d2c-691df43b8892-kube-api-access-285wq\") pod \"ironic-85b58c7f48-9dgpg\" (UID: \"f8ec340a-d11c-4b3b-9d2c-691df43b8892\") " pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.974258 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f8ec340a-d11c-4b3b-9d2c-691df43b8892-config-data-merged\") pod \"ironic-85b58c7f48-9dgpg\" (UID: \"f8ec340a-d11c-4b3b-9d2c-691df43b8892\") " pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.974301 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8ec340a-d11c-4b3b-9d2c-691df43b8892-config-data\") pod \"ironic-85b58c7f48-9dgpg\" (UID: \"f8ec340a-d11c-4b3b-9d2c-691df43b8892\") " pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.974328 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8ec340a-d11c-4b3b-9d2c-691df43b8892-scripts\") pod \"ironic-85b58c7f48-9dgpg\" (UID: \"f8ec340a-d11c-4b3b-9d2c-691df43b8892\") " pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.974387 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8ec340a-d11c-4b3b-9d2c-691df43b8892-config-data-custom\") pod \"ironic-85b58c7f48-9dgpg\" (UID: \"f8ec340a-d11c-4b3b-9d2c-691df43b8892\") " pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.974434 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8ec340a-d11c-4b3b-9d2c-691df43b8892-logs\") pod \"ironic-85b58c7f48-9dgpg\" (UID: \"f8ec340a-d11c-4b3b-9d2c-691df43b8892\") " pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.974459 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/f8ec340a-d11c-4b3b-9d2c-691df43b8892-etc-podinfo\") pod \"ironic-85b58c7f48-9dgpg\" (UID: \"f8ec340a-d11c-4b3b-9d2c-691df43b8892\") " pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.974488 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8ec340a-d11c-4b3b-9d2c-691df43b8892-combined-ca-bundle\") pod \"ironic-85b58c7f48-9dgpg\" (UID: \"f8ec340a-d11c-4b3b-9d2c-691df43b8892\") " pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.974555 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8ec340a-d11c-4b3b-9d2c-691df43b8892-public-tls-certs\") pod \"ironic-85b58c7f48-9dgpg\" (UID: \"f8ec340a-d11c-4b3b-9d2c-691df43b8892\") " pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.974587 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8ec340a-d11c-4b3b-9d2c-691df43b8892-internal-tls-certs\") pod \"ironic-85b58c7f48-9dgpg\" (UID: \"f8ec340a-d11c-4b3b-9d2c-691df43b8892\") " pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.974783 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/f8ec340a-d11c-4b3b-9d2c-691df43b8892-config-data-merged\") pod \"ironic-85b58c7f48-9dgpg\" (UID: \"f8ec340a-d11c-4b3b-9d2c-691df43b8892\") " pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.974810 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f8ec340a-d11c-4b3b-9d2c-691df43b8892-logs\") pod \"ironic-85b58c7f48-9dgpg\" (UID: \"f8ec340a-d11c-4b3b-9d2c-691df43b8892\") " pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.980772 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/f8ec340a-d11c-4b3b-9d2c-691df43b8892-config-data-custom\") pod \"ironic-85b58c7f48-9dgpg\" (UID: \"f8ec340a-d11c-4b3b-9d2c-691df43b8892\") " pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.984620 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/f8ec340a-d11c-4b3b-9d2c-691df43b8892-etc-podinfo\") pod \"ironic-85b58c7f48-9dgpg\" (UID: \"f8ec340a-d11c-4b3b-9d2c-691df43b8892\") " pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.984631 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f8ec340a-d11c-4b3b-9d2c-691df43b8892-config-data\") pod \"ironic-85b58c7f48-9dgpg\" (UID: \"f8ec340a-d11c-4b3b-9d2c-691df43b8892\") " pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.985543 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f8ec340a-d11c-4b3b-9d2c-691df43b8892-combined-ca-bundle\") pod \"ironic-85b58c7f48-9dgpg\" (UID: \"f8ec340a-d11c-4b3b-9d2c-691df43b8892\") " pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.986355 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8ec340a-d11c-4b3b-9d2c-691df43b8892-internal-tls-certs\") pod \"ironic-85b58c7f48-9dgpg\" (UID: \"f8ec340a-d11c-4b3b-9d2c-691df43b8892\") " pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.986371 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f8ec340a-d11c-4b3b-9d2c-691df43b8892-scripts\") pod \"ironic-85b58c7f48-9dgpg\" (UID: \"f8ec340a-d11c-4b3b-9d2c-691df43b8892\") " pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.987431 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f8ec340a-d11c-4b3b-9d2c-691df43b8892-public-tls-certs\") pod \"ironic-85b58c7f48-9dgpg\" (UID: \"f8ec340a-d11c-4b3b-9d2c-691df43b8892\") " pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:46 crc kubenswrapper[4967]: I1011 04:08:46.989563 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-285wq\" (UniqueName: \"kubernetes.io/projected/f8ec340a-d11c-4b3b-9d2c-691df43b8892-kube-api-access-285wq\") pod \"ironic-85b58c7f48-9dgpg\" (UID: \"f8ec340a-d11c-4b3b-9d2c-691df43b8892\") " pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:47 crc kubenswrapper[4967]: I1011 04:08:47.098448 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:47 crc kubenswrapper[4967]: I1011 04:08:47.351315 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-create-bprvt" Oct 11 04:08:47 crc kubenswrapper[4967]: I1011 04:08:47.484177 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dqcl7\" (UniqueName: \"kubernetes.io/projected/12c05ddc-4b39-4924-924c-955b8639a839-kube-api-access-dqcl7\") pod \"12c05ddc-4b39-4924-924c-955b8639a839\" (UID: \"12c05ddc-4b39-4924-924c-955b8639a839\") " Oct 11 04:08:47 crc kubenswrapper[4967]: I1011 04:08:47.489404 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/12c05ddc-4b39-4924-924c-955b8639a839-kube-api-access-dqcl7" (OuterVolumeSpecName: "kube-api-access-dqcl7") pod "12c05ddc-4b39-4924-924c-955b8639a839" (UID: "12c05ddc-4b39-4924-924c-955b8639a839"). InnerVolumeSpecName "kube-api-access-dqcl7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:08:47 crc kubenswrapper[4967]: I1011 04:08:47.585286 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-85b58c7f48-9dgpg"] Oct 11 04:08:47 crc kubenswrapper[4967]: I1011 04:08:47.586459 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dqcl7\" (UniqueName: \"kubernetes.io/projected/12c05ddc-4b39-4924-924c-955b8639a839-kube-api-access-dqcl7\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:47 crc kubenswrapper[4967]: W1011 04:08:47.624422 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf8ec340a_d11c_4b3b_9d2c_691df43b8892.slice/crio-824efc2e568342f673f39580560a3e0ad1da5aaa52148174b0364ba5b02138c6 WatchSource:0}: Error finding container 824efc2e568342f673f39580560a3e0ad1da5aaa52148174b0364ba5b02138c6: Status 404 returned error can't find the container with id 824efc2e568342f673f39580560a3e0ad1da5aaa52148174b0364ba5b02138c6 Oct 11 04:08:47 crc kubenswrapper[4967]: I1011 04:08:47.947217 4967 generic.go:334] "Generic (PLEG): container finished" podID="64c75ea6-aed6-4bf6-8eef-39e76384e14f" containerID="62a3e33ddd46e7b80643f3a4c852455565dfc189794219113986f6815eb3264b" exitCode=0 Oct 11 04:08:47 crc kubenswrapper[4967]: I1011 04:08:47.947274 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"64c75ea6-aed6-4bf6-8eef-39e76384e14f","Type":"ContainerDied","Data":"62a3e33ddd46e7b80643f3a4c852455565dfc189794219113986f6815eb3264b"} Oct 11 04:08:47 crc kubenswrapper[4967]: I1011 04:08:47.958278 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c900849f-54f8-44ed-b759-ae6fbbe441da","Type":"ContainerStarted","Data":"1dfc90218ba03842240e0ca2ef0bfa210917a12d817d66c8abf6ad04bb7eb955"} Oct 11 04:08:47 crc kubenswrapper[4967]: I1011 04:08:47.960054 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-85b58c7f48-9dgpg" event={"ID":"f8ec340a-d11c-4b3b-9d2c-691df43b8892","Type":"ContainerStarted","Data":"824efc2e568342f673f39580560a3e0ad1da5aaa52148174b0364ba5b02138c6"} Oct 11 04:08:47 crc kubenswrapper[4967]: I1011 04:08:47.964246 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-create-bprvt" event={"ID":"12c05ddc-4b39-4924-924c-955b8639a839","Type":"ContainerDied","Data":"df3d198f9ce2a0e29b4ac644d8045c0310c19106ada7e2dc3de68965b39c7de5"} Oct 11 04:08:47 crc kubenswrapper[4967]: I1011 04:08:47.964286 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="df3d198f9ce2a0e29b4ac644d8045c0310c19106ada7e2dc3de68965b39c7de5" Oct 11 04:08:47 crc kubenswrapper[4967]: I1011 04:08:47.964339 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-create-bprvt" Oct 11 04:08:48 crc kubenswrapper[4967]: I1011 04:08:48.244394 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-6554fc45b4-x7btb" Oct 11 04:08:48 crc kubenswrapper[4967]: I1011 04:08:48.981682 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" event={"ID":"4390a950-7185-4fc0-b152-b5927940032b","Type":"ContainerStarted","Data":"35a87df2ce5c675c15d00b9d84c1bc636b767b23e69655ff02543264611901bd"} Oct 11 04:08:48 crc kubenswrapper[4967]: I1011 04:08:48.982145 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" Oct 11 04:08:48 crc kubenswrapper[4967]: I1011 04:08:48.984441 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-api-0" event={"ID":"c900849f-54f8-44ed-b759-ae6fbbe441da","Type":"ContainerStarted","Data":"0eaf96623ca8d6e5ae298d82853735ee3769a3bdf29d570c9346546c94ee5b50"} Oct 11 04:08:48 crc kubenswrapper[4967]: I1011 04:08:48.984569 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/cinder-api-0" Oct 11 04:08:48 crc kubenswrapper[4967]: I1011 04:08:48.992668 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8e8fb63d-44cd-427e-8324-743d83fc8416","Type":"ContainerStarted","Data":"411889abd0b7cb0fda6fd8bb115f5cce070a0f82a59d11ea3ce637cb60dc61c0"} Oct 11 04:08:48 crc kubenswrapper[4967]: I1011 04:08:48.992819 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8e8fb63d-44cd-427e-8324-743d83fc8416" containerName="ceilometer-central-agent" containerID="cri-o://013addb6d98525fab2a9a93cb773a0dd9d6766722ae77ed6790a40a2f9919f9e" gracePeriod=30 Oct 11 04:08:48 crc kubenswrapper[4967]: I1011 04:08:48.993087 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 11 04:08:48 crc kubenswrapper[4967]: I1011 04:08:48.993143 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8e8fb63d-44cd-427e-8324-743d83fc8416" containerName="proxy-httpd" containerID="cri-o://411889abd0b7cb0fda6fd8bb115f5cce070a0f82a59d11ea3ce637cb60dc61c0" gracePeriod=30 Oct 11 04:08:48 crc kubenswrapper[4967]: I1011 04:08:48.993201 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8e8fb63d-44cd-427e-8324-743d83fc8416" containerName="sg-core" containerID="cri-o://5202ac518e2e2e7aaca1d0784961a9f966e120ded550fc4bac7e0bcbbf151e49" gracePeriod=30 Oct 11 04:08:48 crc kubenswrapper[4967]: I1011 04:08:48.993250 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="8e8fb63d-44cd-427e-8324-743d83fc8416" containerName="ceilometer-notification-agent" containerID="cri-o://3021661bce4ef00784088c3e9a64d812c8301954bd3c35ac594364a201d895ee" gracePeriod=30 Oct 11 04:08:49 crc kubenswrapper[4967]: I1011 04:08:49.006935 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" podStartSLOduration=3.060966681 podStartE2EDuration="6.006915146s" podCreationTimestamp="2025-10-11 04:08:43 +0000 UTC" firstStartedPulling="2025-10-11 04:08:45.248354298 +0000 UTC m=+1053.211563231" lastFinishedPulling="2025-10-11 04:08:48.194302763 +0000 UTC m=+1056.157511696" observedRunningTime="2025-10-11 04:08:49.001292987 +0000 UTC m=+1056.964501930" watchObservedRunningTime="2025-10-11 04:08:49.006915146 +0000 UTC m=+1056.970124079" Oct 11 04:08:49 crc kubenswrapper[4967]: I1011 04:08:49.031777 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=8.689253974 podStartE2EDuration="14.031759877s" podCreationTimestamp="2025-10-11 04:08:35 +0000 UTC" firstStartedPulling="2025-10-11 04:08:42.309099573 +0000 UTC m=+1050.272308506" lastFinishedPulling="2025-10-11 04:08:47.651605476 +0000 UTC m=+1055.614814409" observedRunningTime="2025-10-11 04:08:49.027328278 +0000 UTC m=+1056.990537211" watchObservedRunningTime="2025-10-11 04:08:49.031759877 +0000 UTC m=+1056.994968810" Oct 11 04:08:49 crc kubenswrapper[4967]: I1011 04:08:49.048876 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-api-0" podStartSLOduration=4.04886087 podStartE2EDuration="4.04886087s" podCreationTimestamp="2025-10-11 04:08:45 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:08:49.046131422 +0000 UTC m=+1057.009340355" watchObservedRunningTime="2025-10-11 04:08:49.04886087 +0000 UTC m=+1057.012069803" Oct 11 04:08:49 crc kubenswrapper[4967]: I1011 04:08:49.331918 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 11 04:08:49 crc kubenswrapper[4967]: I1011 04:08:49.383159 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 11 04:08:50 crc kubenswrapper[4967]: I1011 04:08:50.017711 4967 generic.go:334] "Generic (PLEG): container finished" podID="8e8fb63d-44cd-427e-8324-743d83fc8416" containerID="411889abd0b7cb0fda6fd8bb115f5cce070a0f82a59d11ea3ce637cb60dc61c0" exitCode=0 Oct 11 04:08:50 crc kubenswrapper[4967]: I1011 04:08:50.018023 4967 generic.go:334] "Generic (PLEG): container finished" podID="8e8fb63d-44cd-427e-8324-743d83fc8416" containerID="5202ac518e2e2e7aaca1d0784961a9f966e120ded550fc4bac7e0bcbbf151e49" exitCode=2 Oct 11 04:08:50 crc kubenswrapper[4967]: I1011 04:08:50.018034 4967 generic.go:334] "Generic (PLEG): container finished" podID="8e8fb63d-44cd-427e-8324-743d83fc8416" containerID="3021661bce4ef00784088c3e9a64d812c8301954bd3c35ac594364a201d895ee" exitCode=0 Oct 11 04:08:50 crc kubenswrapper[4967]: I1011 04:08:50.018043 4967 generic.go:334] "Generic (PLEG): container finished" podID="8e8fb63d-44cd-427e-8324-743d83fc8416" containerID="013addb6d98525fab2a9a93cb773a0dd9d6766722ae77ed6790a40a2f9919f9e" exitCode=0 Oct 11 04:08:50 crc kubenswrapper[4967]: I1011 04:08:50.017787 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8e8fb63d-44cd-427e-8324-743d83fc8416","Type":"ContainerDied","Data":"411889abd0b7cb0fda6fd8bb115f5cce070a0f82a59d11ea3ce637cb60dc61c0"} Oct 11 04:08:50 crc kubenswrapper[4967]: I1011 04:08:50.018376 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8e8fb63d-44cd-427e-8324-743d83fc8416","Type":"ContainerDied","Data":"5202ac518e2e2e7aaca1d0784961a9f966e120ded550fc4bac7e0bcbbf151e49"} Oct 11 04:08:50 crc kubenswrapper[4967]: I1011 04:08:50.018397 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8e8fb63d-44cd-427e-8324-743d83fc8416","Type":"ContainerDied","Data":"3021661bce4ef00784088c3e9a64d812c8301954bd3c35ac594364a201d895ee"} Oct 11 04:08:50 crc kubenswrapper[4967]: I1011 04:08:50.018411 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8e8fb63d-44cd-427e-8324-743d83fc8416","Type":"ContainerDied","Data":"013addb6d98525fab2a9a93cb773a0dd9d6766722ae77ed6790a40a2f9919f9e"} Oct 11 04:08:50 crc kubenswrapper[4967]: I1011 04:08:50.018935 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="4bdd15be-7524-4447-bab5-43c0a6b40cf0" containerName="cinder-scheduler" containerID="cri-o://89575635b09e4fd75c0a08808d714ea4dee6937997a73292f38f9e45e2e88460" gracePeriod=30 Oct 11 04:08:50 crc kubenswrapper[4967]: I1011 04:08:50.019178 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/cinder-scheduler-0" podUID="4bdd15be-7524-4447-bab5-43c0a6b40cf0" containerName="probe" containerID="cri-o://3c1ad1c69eef5590d20f4be3aaf48cd73f8aff961e369759999c3c0b227a7026" gracePeriod=30 Oct 11 04:08:50 crc kubenswrapper[4967]: I1011 04:08:50.039362 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:08:50 crc kubenswrapper[4967]: I1011 04:08:50.132965 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8e8fb63d-44cd-427e-8324-743d83fc8416-run-httpd\") pod \"8e8fb63d-44cd-427e-8324-743d83fc8416\" (UID: \"8e8fb63d-44cd-427e-8324-743d83fc8416\") " Oct 11 04:08:50 crc kubenswrapper[4967]: I1011 04:08:50.133019 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e8fb63d-44cd-427e-8324-743d83fc8416-config-data\") pod \"8e8fb63d-44cd-427e-8324-743d83fc8416\" (UID: \"8e8fb63d-44cd-427e-8324-743d83fc8416\") " Oct 11 04:08:50 crc kubenswrapper[4967]: I1011 04:08:50.133052 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bvstw\" (UniqueName: \"kubernetes.io/projected/8e8fb63d-44cd-427e-8324-743d83fc8416-kube-api-access-bvstw\") pod \"8e8fb63d-44cd-427e-8324-743d83fc8416\" (UID: \"8e8fb63d-44cd-427e-8324-743d83fc8416\") " Oct 11 04:08:50 crc kubenswrapper[4967]: I1011 04:08:50.133151 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e8fb63d-44cd-427e-8324-743d83fc8416-combined-ca-bundle\") pod \"8e8fb63d-44cd-427e-8324-743d83fc8416\" (UID: \"8e8fb63d-44cd-427e-8324-743d83fc8416\") " Oct 11 04:08:50 crc kubenswrapper[4967]: I1011 04:08:50.133182 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8e8fb63d-44cd-427e-8324-743d83fc8416-log-httpd\") pod \"8e8fb63d-44cd-427e-8324-743d83fc8416\" (UID: \"8e8fb63d-44cd-427e-8324-743d83fc8416\") " Oct 11 04:08:50 crc kubenswrapper[4967]: I1011 04:08:50.133283 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e8fb63d-44cd-427e-8324-743d83fc8416-scripts\") pod \"8e8fb63d-44cd-427e-8324-743d83fc8416\" (UID: \"8e8fb63d-44cd-427e-8324-743d83fc8416\") " Oct 11 04:08:50 crc kubenswrapper[4967]: I1011 04:08:50.133341 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8e8fb63d-44cd-427e-8324-743d83fc8416-sg-core-conf-yaml\") pod \"8e8fb63d-44cd-427e-8324-743d83fc8416\" (UID: \"8e8fb63d-44cd-427e-8324-743d83fc8416\") " Oct 11 04:08:50 crc kubenswrapper[4967]: I1011 04:08:50.133640 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e8fb63d-44cd-427e-8324-743d83fc8416-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "8e8fb63d-44cd-427e-8324-743d83fc8416" (UID: "8e8fb63d-44cd-427e-8324-743d83fc8416"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:08:50 crc kubenswrapper[4967]: I1011 04:08:50.133914 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8e8fb63d-44cd-427e-8324-743d83fc8416-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "8e8fb63d-44cd-427e-8324-743d83fc8416" (UID: "8e8fb63d-44cd-427e-8324-743d83fc8416"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:08:50 crc kubenswrapper[4967]: I1011 04:08:50.134054 4967 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8e8fb63d-44cd-427e-8324-743d83fc8416-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:50 crc kubenswrapper[4967]: I1011 04:08:50.134106 4967 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/8e8fb63d-44cd-427e-8324-743d83fc8416-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:50 crc kubenswrapper[4967]: I1011 04:08:50.137954 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8e8fb63d-44cd-427e-8324-743d83fc8416-kube-api-access-bvstw" (OuterVolumeSpecName: "kube-api-access-bvstw") pod "8e8fb63d-44cd-427e-8324-743d83fc8416" (UID: "8e8fb63d-44cd-427e-8324-743d83fc8416"). InnerVolumeSpecName "kube-api-access-bvstw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:08:50 crc kubenswrapper[4967]: I1011 04:08:50.139238 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e8fb63d-44cd-427e-8324-743d83fc8416-scripts" (OuterVolumeSpecName: "scripts") pod "8e8fb63d-44cd-427e-8324-743d83fc8416" (UID: "8e8fb63d-44cd-427e-8324-743d83fc8416"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:50 crc kubenswrapper[4967]: I1011 04:08:50.167799 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e8fb63d-44cd-427e-8324-743d83fc8416-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "8e8fb63d-44cd-427e-8324-743d83fc8416" (UID: "8e8fb63d-44cd-427e-8324-743d83fc8416"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:50 crc kubenswrapper[4967]: I1011 04:08:50.219275 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e8fb63d-44cd-427e-8324-743d83fc8416-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "8e8fb63d-44cd-427e-8324-743d83fc8416" (UID: "8e8fb63d-44cd-427e-8324-743d83fc8416"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:50 crc kubenswrapper[4967]: I1011 04:08:50.234228 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8e8fb63d-44cd-427e-8324-743d83fc8416-config-data" (OuterVolumeSpecName: "config-data") pod "8e8fb63d-44cd-427e-8324-743d83fc8416" (UID: "8e8fb63d-44cd-427e-8324-743d83fc8416"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:50 crc kubenswrapper[4967]: I1011 04:08:50.235292 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e8fb63d-44cd-427e-8324-743d83fc8416-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:50 crc kubenswrapper[4967]: I1011 04:08:50.235312 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bvstw\" (UniqueName: \"kubernetes.io/projected/8e8fb63d-44cd-427e-8324-743d83fc8416-kube-api-access-bvstw\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:50 crc kubenswrapper[4967]: I1011 04:08:50.235324 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e8fb63d-44cd-427e-8324-743d83fc8416-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:50 crc kubenswrapper[4967]: I1011 04:08:50.235333 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/8e8fb63d-44cd-427e-8324-743d83fc8416-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:50 crc kubenswrapper[4967]: I1011 04:08:50.235340 4967 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/8e8fb63d-44cd-427e-8324-743d83fc8416-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.044099 4967 generic.go:334] "Generic (PLEG): container finished" podID="5c82b179-3510-4fb2-9352-d41e7193d221" containerID="fab633c4576fe35b347aa6b382b58410ac65fa121b618fac8f93e6338a61127b" exitCode=0 Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.044265 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-5b5f75f756-jlhwg" event={"ID":"5c82b179-3510-4fb2-9352-d41e7193d221","Type":"ContainerDied","Data":"fab633c4576fe35b347aa6b382b58410ac65fa121b618fac8f93e6338a61127b"} Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.051529 4967 generic.go:334] "Generic (PLEG): container finished" podID="4bdd15be-7524-4447-bab5-43c0a6b40cf0" containerID="3c1ad1c69eef5590d20f4be3aaf48cd73f8aff961e369759999c3c0b227a7026" exitCode=0 Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.051606 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4bdd15be-7524-4447-bab5-43c0a6b40cf0","Type":"ContainerDied","Data":"3c1ad1c69eef5590d20f4be3aaf48cd73f8aff961e369759999c3c0b227a7026"} Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.056355 4967 generic.go:334] "Generic (PLEG): container finished" podID="f8ec340a-d11c-4b3b-9d2c-691df43b8892" containerID="589dd019079f8bf28ea84014750734f58f94c3b6feb08ac68f694881d7ef5c65" exitCode=0 Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.057311 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-85b58c7f48-9dgpg" event={"ID":"f8ec340a-d11c-4b3b-9d2c-691df43b8892","Type":"ContainerDied","Data":"589dd019079f8bf28ea84014750734f58f94c3b6feb08ac68f694881d7ef5c65"} Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.062327 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"8e8fb63d-44cd-427e-8324-743d83fc8416","Type":"ContainerDied","Data":"f21cdca371f1bb647b12cf6cb96e7d808e5322b3a6d444b6285a714cf3b2f7b0"} Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.062385 4967 scope.go:117] "RemoveContainer" containerID="411889abd0b7cb0fda6fd8bb115f5cce070a0f82a59d11ea3ce637cb60dc61c0" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.062536 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.119217 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.135651 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.141880 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:08:51 crc kubenswrapper[4967]: E1011 04:08:51.142310 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e8fb63d-44cd-427e-8324-743d83fc8416" containerName="proxy-httpd" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.142324 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e8fb63d-44cd-427e-8324-743d83fc8416" containerName="proxy-httpd" Oct 11 04:08:51 crc kubenswrapper[4967]: E1011 04:08:51.142340 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e8fb63d-44cd-427e-8324-743d83fc8416" containerName="ceilometer-notification-agent" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.142345 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e8fb63d-44cd-427e-8324-743d83fc8416" containerName="ceilometer-notification-agent" Oct 11 04:08:51 crc kubenswrapper[4967]: E1011 04:08:51.142369 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="12c05ddc-4b39-4924-924c-955b8639a839" containerName="mariadb-database-create" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.142375 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="12c05ddc-4b39-4924-924c-955b8639a839" containerName="mariadb-database-create" Oct 11 04:08:51 crc kubenswrapper[4967]: E1011 04:08:51.142606 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e8fb63d-44cd-427e-8324-743d83fc8416" containerName="sg-core" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.142618 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e8fb63d-44cd-427e-8324-743d83fc8416" containerName="sg-core" Oct 11 04:08:51 crc kubenswrapper[4967]: E1011 04:08:51.142628 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8e8fb63d-44cd-427e-8324-743d83fc8416" containerName="ceilometer-central-agent" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.142634 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="8e8fb63d-44cd-427e-8324-743d83fc8416" containerName="ceilometer-central-agent" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.142800 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e8fb63d-44cd-427e-8324-743d83fc8416" containerName="proxy-httpd" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.142842 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e8fb63d-44cd-427e-8324-743d83fc8416" containerName="ceilometer-central-agent" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.142852 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="12c05ddc-4b39-4924-924c-955b8639a839" containerName="mariadb-database-create" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.142877 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e8fb63d-44cd-427e-8324-743d83fc8416" containerName="ceilometer-notification-agent" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.142888 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="8e8fb63d-44cd-427e-8324-743d83fc8416" containerName="sg-core" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.144551 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.145888 4967 scope.go:117] "RemoveContainer" containerID="5202ac518e2e2e7aaca1d0784961a9f966e120ded550fc4bac7e0bcbbf151e49" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.146835 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.146863 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.150272 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.169459 4967 scope.go:117] "RemoveContainer" containerID="3021661bce4ef00784088c3e9a64d812c8301954bd3c35ac594364a201d895ee" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.249225 4967 scope.go:117] "RemoveContainer" containerID="013addb6d98525fab2a9a93cb773a0dd9d6766722ae77ed6790a40a2f9919f9e" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.257318 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\") " pod="openstack/ceilometer-0" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.257435 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-config-data\") pod \"ceilometer-0\" (UID: \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\") " pod="openstack/ceilometer-0" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.257455 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-log-httpd\") pod \"ceilometer-0\" (UID: \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\") " pod="openstack/ceilometer-0" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.257640 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dh476\" (UniqueName: \"kubernetes.io/projected/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-kube-api-access-dh476\") pod \"ceilometer-0\" (UID: \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\") " pod="openstack/ceilometer-0" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.257744 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\") " pod="openstack/ceilometer-0" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.257818 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-scripts\") pod \"ceilometer-0\" (UID: \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\") " pod="openstack/ceilometer-0" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.257858 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-run-httpd\") pod \"ceilometer-0\" (UID: \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\") " pod="openstack/ceilometer-0" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.359898 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-config-data\") pod \"ceilometer-0\" (UID: \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\") " pod="openstack/ceilometer-0" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.360144 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-log-httpd\") pod \"ceilometer-0\" (UID: \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\") " pod="openstack/ceilometer-0" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.360201 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dh476\" (UniqueName: \"kubernetes.io/projected/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-kube-api-access-dh476\") pod \"ceilometer-0\" (UID: \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\") " pod="openstack/ceilometer-0" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.360233 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\") " pod="openstack/ceilometer-0" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.360277 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-scripts\") pod \"ceilometer-0\" (UID: \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\") " pod="openstack/ceilometer-0" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.360303 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-run-httpd\") pod \"ceilometer-0\" (UID: \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\") " pod="openstack/ceilometer-0" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.360338 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\") " pod="openstack/ceilometer-0" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.363641 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-log-httpd\") pod \"ceilometer-0\" (UID: \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\") " pod="openstack/ceilometer-0" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.364711 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-run-httpd\") pod \"ceilometer-0\" (UID: \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\") " pod="openstack/ceilometer-0" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.365866 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\") " pod="openstack/ceilometer-0" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.367775 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-config-data\") pod \"ceilometer-0\" (UID: \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\") " pod="openstack/ceilometer-0" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.375531 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\") " pod="openstack/ceilometer-0" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.383598 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-scripts\") pod \"ceilometer-0\" (UID: \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\") " pod="openstack/ceilometer-0" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.387649 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dh476\" (UniqueName: \"kubernetes.io/projected/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-kube-api-access-dh476\") pod \"ceilometer-0\" (UID: \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\") " pod="openstack/ceilometer-0" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.469242 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:08:51 crc kubenswrapper[4967]: I1011 04:08:51.891564 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:08:52 crc kubenswrapper[4967]: I1011 04:08:52.052295 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/neutron-5f96fd6969-6srgk" Oct 11 04:08:52 crc kubenswrapper[4967]: I1011 04:08:52.075095 4967 generic.go:334] "Generic (PLEG): container finished" podID="4390a950-7185-4fc0-b152-b5927940032b" containerID="35a87df2ce5c675c15d00b9d84c1bc636b767b23e69655ff02543264611901bd" exitCode=1 Oct 11 04:08:52 crc kubenswrapper[4967]: I1011 04:08:52.075173 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" event={"ID":"4390a950-7185-4fc0-b152-b5927940032b","Type":"ContainerDied","Data":"35a87df2ce5c675c15d00b9d84c1bc636b767b23e69655ff02543264611901bd"} Oct 11 04:08:52 crc kubenswrapper[4967]: I1011 04:08:52.076592 4967 scope.go:117] "RemoveContainer" containerID="35a87df2ce5c675c15d00b9d84c1bc636b767b23e69655ff02543264611901bd" Oct 11 04:08:52 crc kubenswrapper[4967]: I1011 04:08:52.088934 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-85b58c7f48-9dgpg" event={"ID":"f8ec340a-d11c-4b3b-9d2c-691df43b8892","Type":"ContainerStarted","Data":"a8f302a1123ccee8c4279d217c8e06e3756d02e28799253d47efd74323a00292"} Oct 11 04:08:52 crc kubenswrapper[4967]: I1011 04:08:52.102863 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-5b5f75f756-jlhwg" event={"ID":"5c82b179-3510-4fb2-9352-d41e7193d221","Type":"ContainerStarted","Data":"565f03d42b75e92dcace4c1ad5111a96bc6b2d2fbcbc46bc88b7137676ce22f6"} Oct 11 04:08:52 crc kubenswrapper[4967]: I1011 04:08:52.144000 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6554fc45b4-x7btb"] Oct 11 04:08:52 crc kubenswrapper[4967]: I1011 04:08:52.144466 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6554fc45b4-x7btb" podUID="92205841-5c2e-4fc7-9328-8a164f5fb8cf" containerName="neutron-api" containerID="cri-o://97c51df86213c4cd58189937091722374bf529014675fdfdb2cbbc6d2816f6dd" gracePeriod=30 Oct 11 04:08:52 crc kubenswrapper[4967]: I1011 04:08:52.144527 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/neutron-6554fc45b4-x7btb" podUID="92205841-5c2e-4fc7-9328-8a164f5fb8cf" containerName="neutron-httpd" containerID="cri-o://50f50943493197ea90b614621946c4164887b9cfb4c14fbe86b0b763dd2f42b6" gracePeriod=30 Oct 11 04:08:52 crc kubenswrapper[4967]: I1011 04:08:52.621681 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-db-create-zb7zk"] Oct 11 04:08:52 crc kubenswrapper[4967]: I1011 04:08:52.622816 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-zb7zk" Oct 11 04:08:52 crc kubenswrapper[4967]: I1011 04:08:52.644686 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-zb7zk"] Oct 11 04:08:52 crc kubenswrapper[4967]: I1011 04:08:52.689054 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mrqsh\" (UniqueName: \"kubernetes.io/projected/706a95a7-728b-40e8-8c4e-221cc836295f-kube-api-access-mrqsh\") pod \"nova-api-db-create-zb7zk\" (UID: \"706a95a7-728b-40e8-8c4e-221cc836295f\") " pod="openstack/nova-api-db-create-zb7zk" Oct 11 04:08:52 crc kubenswrapper[4967]: I1011 04:08:52.735214 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-db-create-zt7cl"] Oct 11 04:08:52 crc kubenswrapper[4967]: I1011 04:08:52.741889 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-zt7cl" Oct 11 04:08:52 crc kubenswrapper[4967]: I1011 04:08:52.755545 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-zt7cl"] Oct 11 04:08:52 crc kubenswrapper[4967]: I1011 04:08:52.791772 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dtsq\" (UniqueName: \"kubernetes.io/projected/540b1dfa-8e24-4aed-a3d9-6c2a1e367155-kube-api-access-6dtsq\") pod \"nova-cell0-db-create-zt7cl\" (UID: \"540b1dfa-8e24-4aed-a3d9-6c2a1e367155\") " pod="openstack/nova-cell0-db-create-zt7cl" Oct 11 04:08:52 crc kubenswrapper[4967]: I1011 04:08:52.791862 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mrqsh\" (UniqueName: \"kubernetes.io/projected/706a95a7-728b-40e8-8c4e-221cc836295f-kube-api-access-mrqsh\") pod \"nova-api-db-create-zb7zk\" (UID: \"706a95a7-728b-40e8-8c4e-221cc836295f\") " pod="openstack/nova-api-db-create-zb7zk" Oct 11 04:08:52 crc kubenswrapper[4967]: I1011 04:08:52.825154 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mrqsh\" (UniqueName: \"kubernetes.io/projected/706a95a7-728b-40e8-8c4e-221cc836295f-kube-api-access-mrqsh\") pod \"nova-api-db-create-zb7zk\" (UID: \"706a95a7-728b-40e8-8c4e-221cc836295f\") " pod="openstack/nova-api-db-create-zb7zk" Oct 11 04:08:52 crc kubenswrapper[4967]: I1011 04:08:52.833878 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8e8fb63d-44cd-427e-8324-743d83fc8416" path="/var/lib/kubelet/pods/8e8fb63d-44cd-427e-8324-743d83fc8416/volumes" Oct 11 04:08:52 crc kubenswrapper[4967]: I1011 04:08:52.834731 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-db-create-zth24"] Oct 11 04:08:52 crc kubenswrapper[4967]: I1011 04:08:52.838768 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-zth24" Oct 11 04:08:52 crc kubenswrapper[4967]: I1011 04:08:52.840286 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-zth24"] Oct 11 04:08:52 crc kubenswrapper[4967]: I1011 04:08:52.905662 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dtsq\" (UniqueName: \"kubernetes.io/projected/540b1dfa-8e24-4aed-a3d9-6c2a1e367155-kube-api-access-6dtsq\") pod \"nova-cell0-db-create-zt7cl\" (UID: \"540b1dfa-8e24-4aed-a3d9-6c2a1e367155\") " pod="openstack/nova-cell0-db-create-zt7cl" Oct 11 04:08:52 crc kubenswrapper[4967]: I1011 04:08:52.906428 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x4srh\" (UniqueName: \"kubernetes.io/projected/098dc82c-7d36-48e4-bc48-28e8721f6617-kube-api-access-x4srh\") pod \"nova-cell1-db-create-zth24\" (UID: \"098dc82c-7d36-48e4-bc48-28e8721f6617\") " pod="openstack/nova-cell1-db-create-zth24" Oct 11 04:08:52 crc kubenswrapper[4967]: I1011 04:08:52.927661 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dtsq\" (UniqueName: \"kubernetes.io/projected/540b1dfa-8e24-4aed-a3d9-6c2a1e367155-kube-api-access-6dtsq\") pod \"nova-cell0-db-create-zt7cl\" (UID: \"540b1dfa-8e24-4aed-a3d9-6c2a1e367155\") " pod="openstack/nova-cell0-db-create-zt7cl" Oct 11 04:08:52 crc kubenswrapper[4967]: I1011 04:08:52.976798 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-zb7zk" Oct 11 04:08:53 crc kubenswrapper[4967]: I1011 04:08:53.007712 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x4srh\" (UniqueName: \"kubernetes.io/projected/098dc82c-7d36-48e4-bc48-28e8721f6617-kube-api-access-x4srh\") pod \"nova-cell1-db-create-zth24\" (UID: \"098dc82c-7d36-48e4-bc48-28e8721f6617\") " pod="openstack/nova-cell1-db-create-zth24" Oct 11 04:08:53 crc kubenswrapper[4967]: I1011 04:08:53.047179 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x4srh\" (UniqueName: \"kubernetes.io/projected/098dc82c-7d36-48e4-bc48-28e8721f6617-kube-api-access-x4srh\") pod \"nova-cell1-db-create-zth24\" (UID: \"098dc82c-7d36-48e4-bc48-28e8721f6617\") " pod="openstack/nova-cell1-db-create-zth24" Oct 11 04:08:53 crc kubenswrapper[4967]: I1011 04:08:53.076697 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-zt7cl" Oct 11 04:08:53 crc kubenswrapper[4967]: I1011 04:08:53.113845 4967 generic.go:334] "Generic (PLEG): container finished" podID="92205841-5c2e-4fc7-9328-8a164f5fb8cf" containerID="50f50943493197ea90b614621946c4164887b9cfb4c14fbe86b0b763dd2f42b6" exitCode=0 Oct 11 04:08:53 crc kubenswrapper[4967]: I1011 04:08:53.113911 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6554fc45b4-x7btb" event={"ID":"92205841-5c2e-4fc7-9328-8a164f5fb8cf","Type":"ContainerDied","Data":"50f50943493197ea90b614621946c4164887b9cfb4c14fbe86b0b763dd2f42b6"} Oct 11 04:08:53 crc kubenswrapper[4967]: I1011 04:08:53.118483 4967 generic.go:334] "Generic (PLEG): container finished" podID="4bdd15be-7524-4447-bab5-43c0a6b40cf0" containerID="89575635b09e4fd75c0a08808d714ea4dee6937997a73292f38f9e45e2e88460" exitCode=0 Oct 11 04:08:53 crc kubenswrapper[4967]: I1011 04:08:53.118543 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4bdd15be-7524-4447-bab5-43c0a6b40cf0","Type":"ContainerDied","Data":"89575635b09e4fd75c0a08808d714ea4dee6937997a73292f38f9e45e2e88460"} Oct 11 04:08:53 crc kubenswrapper[4967]: I1011 04:08:53.223559 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-zth24" Oct 11 04:08:53 crc kubenswrapper[4967]: I1011 04:08:53.379681 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:08:53 crc kubenswrapper[4967]: I1011 04:08:53.629886 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 11 04:08:53 crc kubenswrapper[4967]: I1011 04:08:53.630358 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f3fcbf92-aff2-444c-abf1-b23c45ef96e9" containerName="glance-log" containerID="cri-o://0e68c39cb6ce4b2246948b7da0836b7edf0d64a7d6f59868bf31e93d2857f817" gracePeriod=30 Oct 11 04:08:53 crc kubenswrapper[4967]: I1011 04:08:53.630753 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-external-api-0" podUID="f3fcbf92-aff2-444c-abf1-b23c45ef96e9" containerName="glance-httpd" containerID="cri-o://22aca31c65aece17da7d51c55636940bceadde7fd9c8f3b5c5066057250f4000" gracePeriod=30 Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.009052 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-inspector-ba91-account-create-595xr"] Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.010263 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-ba91-account-create-595xr" Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.016408 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-db-secret" Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.023898 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-ba91-account-create-595xr"] Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.134130 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gmqzl\" (UniqueName: \"kubernetes.io/projected/de9db0f0-79dd-4edc-8022-c32a3a1340a1-kube-api-access-gmqzl\") pod \"ironic-inspector-ba91-account-create-595xr\" (UID: \"de9db0f0-79dd-4edc-8022-c32a3a1340a1\") " pod="openstack/ironic-inspector-ba91-account-create-595xr" Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.137890 4967 generic.go:334] "Generic (PLEG): container finished" podID="f3fcbf92-aff2-444c-abf1-b23c45ef96e9" containerID="0e68c39cb6ce4b2246948b7da0836b7edf0d64a7d6f59868bf31e93d2857f817" exitCode=143 Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.137953 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f3fcbf92-aff2-444c-abf1-b23c45ef96e9","Type":"ContainerDied","Data":"0e68c39cb6ce4b2246948b7da0836b7edf0d64a7d6f59868bf31e93d2857f817"} Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.235760 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gmqzl\" (UniqueName: \"kubernetes.io/projected/de9db0f0-79dd-4edc-8022-c32a3a1340a1-kube-api-access-gmqzl\") pod \"ironic-inspector-ba91-account-create-595xr\" (UID: \"de9db0f0-79dd-4edc-8022-c32a3a1340a1\") " pod="openstack/ironic-inspector-ba91-account-create-595xr" Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.258336 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gmqzl\" (UniqueName: \"kubernetes.io/projected/de9db0f0-79dd-4edc-8022-c32a3a1340a1-kube-api-access-gmqzl\") pod \"ironic-inspector-ba91-account-create-595xr\" (UID: \"de9db0f0-79dd-4edc-8022-c32a3a1340a1\") " pod="openstack/ironic-inspector-ba91-account-create-595xr" Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.333399 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-ba91-account-create-595xr" Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.442943 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.643862 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.644138 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d8df67d1-7a02-4e18-932a-433412376643" containerName="glance-log" containerID="cri-o://223177e0c8f9782ee3f095403a0314a51431ed701e59885aa682f2114dd478b2" gracePeriod=30 Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.644242 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/glance-default-internal-api-0" podUID="d8df67d1-7a02-4e18-932a-433412376643" containerName="glance-httpd" containerID="cri-o://78ddb9d89814b4e9e287119e168712ef35d876c9d62bc9a8ead0ec4db0ae1f75" gracePeriod=30 Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.696907 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.846911 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4bdd15be-7524-4447-bab5-43c0a6b40cf0-scripts\") pod \"4bdd15be-7524-4447-bab5-43c0a6b40cf0\" (UID: \"4bdd15be-7524-4447-bab5-43c0a6b40cf0\") " Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.847284 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4bdd15be-7524-4447-bab5-43c0a6b40cf0-etc-machine-id\") pod \"4bdd15be-7524-4447-bab5-43c0a6b40cf0\" (UID: \"4bdd15be-7524-4447-bab5-43c0a6b40cf0\") " Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.847465 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4bdd15be-7524-4447-bab5-43c0a6b40cf0-config-data-custom\") pod \"4bdd15be-7524-4447-bab5-43c0a6b40cf0\" (UID: \"4bdd15be-7524-4447-bab5-43c0a6b40cf0\") " Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.847496 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/4bdd15be-7524-4447-bab5-43c0a6b40cf0-etc-machine-id" (OuterVolumeSpecName: "etc-machine-id") pod "4bdd15be-7524-4447-bab5-43c0a6b40cf0" (UID: "4bdd15be-7524-4447-bab5-43c0a6b40cf0"). InnerVolumeSpecName "etc-machine-id". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.847511 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-svs7q\" (UniqueName: \"kubernetes.io/projected/4bdd15be-7524-4447-bab5-43c0a6b40cf0-kube-api-access-svs7q\") pod \"4bdd15be-7524-4447-bab5-43c0a6b40cf0\" (UID: \"4bdd15be-7524-4447-bab5-43c0a6b40cf0\") " Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.847613 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bdd15be-7524-4447-bab5-43c0a6b40cf0-combined-ca-bundle\") pod \"4bdd15be-7524-4447-bab5-43c0a6b40cf0\" (UID: \"4bdd15be-7524-4447-bab5-43c0a6b40cf0\") " Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.847640 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bdd15be-7524-4447-bab5-43c0a6b40cf0-config-data\") pod \"4bdd15be-7524-4447-bab5-43c0a6b40cf0\" (UID: \"4bdd15be-7524-4447-bab5-43c0a6b40cf0\") " Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.848810 4967 reconciler_common.go:293] "Volume detached for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/4bdd15be-7524-4447-bab5-43c0a6b40cf0-etc-machine-id\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.853245 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bdd15be-7524-4447-bab5-43c0a6b40cf0-scripts" (OuterVolumeSpecName: "scripts") pod "4bdd15be-7524-4447-bab5-43c0a6b40cf0" (UID: "4bdd15be-7524-4447-bab5-43c0a6b40cf0"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.854235 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bdd15be-7524-4447-bab5-43c0a6b40cf0-kube-api-access-svs7q" (OuterVolumeSpecName: "kube-api-access-svs7q") pod "4bdd15be-7524-4447-bab5-43c0a6b40cf0" (UID: "4bdd15be-7524-4447-bab5-43c0a6b40cf0"). InnerVolumeSpecName "kube-api-access-svs7q". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.857416 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-db-create-zt7cl"] Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.857751 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bdd15be-7524-4447-bab5-43c0a6b40cf0-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "4bdd15be-7524-4447-bab5-43c0a6b40cf0" (UID: "4bdd15be-7524-4447-bab5-43c0a6b40cf0"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.880110 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-db-create-zb7zk"] Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.891043 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-db-create-zth24"] Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.916470 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bdd15be-7524-4447-bab5-43c0a6b40cf0-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4bdd15be-7524-4447-bab5-43c0a6b40cf0" (UID: "4bdd15be-7524-4447-bab5-43c0a6b40cf0"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.950230 4967 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/4bdd15be-7524-4447-bab5-43c0a6b40cf0-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.950260 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-svs7q\" (UniqueName: \"kubernetes.io/projected/4bdd15be-7524-4447-bab5-43c0a6b40cf0-kube-api-access-svs7q\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.950271 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4bdd15be-7524-4447-bab5-43c0a6b40cf0-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:54 crc kubenswrapper[4967]: I1011 04:08:54.950279 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4bdd15be-7524-4447-bab5-43c0a6b40cf0-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.010150 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4bdd15be-7524-4447-bab5-43c0a6b40cf0-config-data" (OuterVolumeSpecName: "config-data") pod "4bdd15be-7524-4447-bab5-43c0a6b40cf0" (UID: "4bdd15be-7524-4447-bab5-43c0a6b40cf0"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.048649 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-ba91-account-create-595xr"] Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.053485 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4bdd15be-7524-4447-bab5-43c0a6b40cf0-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.157874 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-zth24" event={"ID":"098dc82c-7d36-48e4-bc48-28e8721f6617","Type":"ContainerStarted","Data":"383dd9d4109b281e5b663d9eeb8e34c15b3a8486d8b06d354d56d1399811c062"} Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.159092 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-zth24" event={"ID":"098dc82c-7d36-48e4-bc48-28e8721f6617","Type":"ContainerStarted","Data":"9ce5713be9ef6272f759025c19b4643dddaad0710b4a2482d5b2c879096540b6"} Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.170331 4967 generic.go:334] "Generic (PLEG): container finished" podID="d8df67d1-7a02-4e18-932a-433412376643" containerID="223177e0c8f9782ee3f095403a0314a51431ed701e59885aa682f2114dd478b2" exitCode=143 Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.170391 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d8df67d1-7a02-4e18-932a-433412376643","Type":"ContainerDied","Data":"223177e0c8f9782ee3f095403a0314a51431ed701e59885aa682f2114dd478b2"} Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.174052 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.174530 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"4bdd15be-7524-4447-bab5-43c0a6b40cf0","Type":"ContainerDied","Data":"5eb186429e6c0f24e596aec4d6e354562066fb6d2708ca4aa4ac9a019e2da3e0"} Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.174585 4967 scope.go:117] "RemoveContainer" containerID="3c1ad1c69eef5590d20f4be3aaf48cd73f8aff961e369759999c3c0b227a7026" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.177417 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-db-create-zth24" podStartSLOduration=3.177405196 podStartE2EDuration="3.177405196s" podCreationTimestamp="2025-10-11 04:08:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:08:55.169601577 +0000 UTC m=+1063.132810500" watchObservedRunningTime="2025-10-11 04:08:55.177405196 +0000 UTC m=+1063.140614119" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.187642 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-zb7zk" event={"ID":"706a95a7-728b-40e8-8c4e-221cc836295f","Type":"ContainerStarted","Data":"82f3a6bd6722ea37f904a5fbd419d21d038cfd670bba6e8fa7206aed09197121"} Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.187682 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-zb7zk" event={"ID":"706a95a7-728b-40e8-8c4e-221cc836295f","Type":"ContainerStarted","Data":"6c24146f4bba5945fb8c7ddc6ec6865fc3f1f5b76ad9a0c577864c549c6026cb"} Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.211840 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" event={"ID":"4390a950-7185-4fc0-b152-b5927940032b","Type":"ContainerStarted","Data":"b8c1b1f153783c154d8166fe2c821e9aac2b99008b9620c42dc8206811d3e2aa"} Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.211949 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.215106 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-db-create-zb7zk" podStartSLOduration=3.215088724 podStartE2EDuration="3.215088724s" podCreationTimestamp="2025-10-11 04:08:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:08:55.210147756 +0000 UTC m=+1063.173356699" watchObservedRunningTime="2025-10-11 04:08:55.215088724 +0000 UTC m=+1063.178297647" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.226888 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-85b58c7f48-9dgpg" event={"ID":"f8ec340a-d11c-4b3b-9d2c-691df43b8892","Type":"ContainerStarted","Data":"3be59a1a4107b583eae5b1d33d63566739264b34ff3a95b0d338145839c37c02"} Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.228020 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.241327 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-ba91-account-create-595xr" event={"ID":"de9db0f0-79dd-4edc-8022-c32a3a1340a1","Type":"ContainerStarted","Data":"0bacacd3eff1584a6ee5f64df2350f69a6814867989d5ac49c84a39594ef9e41"} Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.260726 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-zt7cl" event={"ID":"540b1dfa-8e24-4aed-a3d9-6c2a1e367155","Type":"ContainerStarted","Data":"e018e7db55ec5edeecce65b93021dd2c61dd64dd6b73dd0f4ea82c4c455246a7"} Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.260775 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-zt7cl" event={"ID":"540b1dfa-8e24-4aed-a3d9-6c2a1e367155","Type":"ContainerStarted","Data":"7288b0044e5bb280ff8bd2888b6619cf178d21b0daa876a455953cc24ae11355"} Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.267735 4967 generic.go:334] "Generic (PLEG): container finished" podID="5c82b179-3510-4fb2-9352-d41e7193d221" containerID="5ab54b7897056a999ed147d085d6fa895270af2bd90cb31aa6dd4ff5bcf5712b" exitCode=1 Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.267833 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-5b5f75f756-jlhwg" event={"ID":"5c82b179-3510-4fb2-9352-d41e7193d221","Type":"ContainerDied","Data":"5ab54b7897056a999ed147d085d6fa895270af2bd90cb31aa6dd4ff5bcf5712b"} Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.268646 4967 scope.go:117] "RemoveContainer" containerID="5ab54b7897056a999ed147d085d6fa895270af2bd90cb31aa6dd4ff5bcf5712b" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.269123 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eebc2018-61d9-4fe8-ace1-fcb6d817fe80","Type":"ContainerStarted","Data":"20317c8eb3833736a0a77dafb89a49b6926509f49b3a108af90686e6a209a5e8"} Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.277587 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-85b58c7f48-9dgpg" podStartSLOduration=7.256540965 podStartE2EDuration="9.277571013s" podCreationTimestamp="2025-10-11 04:08:46 +0000 UTC" firstStartedPulling="2025-10-11 04:08:47.648645934 +0000 UTC m=+1055.611854867" lastFinishedPulling="2025-10-11 04:08:49.669675982 +0000 UTC m=+1057.632884915" observedRunningTime="2025-10-11 04:08:55.260939768 +0000 UTC m=+1063.224148711" watchObservedRunningTime="2025-10-11 04:08:55.277571013 +0000 UTC m=+1063.240779946" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.431064 4967 scope.go:117] "RemoveContainer" containerID="89575635b09e4fd75c0a08808d714ea4dee6937997a73292f38f9e45e2e88460" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.481797 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.503085 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.516136 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/cinder-scheduler-0"] Oct 11 04:08:55 crc kubenswrapper[4967]: E1011 04:08:55.517445 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bdd15be-7524-4447-bab5-43c0a6b40cf0" containerName="cinder-scheduler" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.517563 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bdd15be-7524-4447-bab5-43c0a6b40cf0" containerName="cinder-scheduler" Oct 11 04:08:55 crc kubenswrapper[4967]: E1011 04:08:55.517645 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4bdd15be-7524-4447-bab5-43c0a6b40cf0" containerName="probe" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.517733 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="4bdd15be-7524-4447-bab5-43c0a6b40cf0" containerName="probe" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.518052 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bdd15be-7524-4447-bab5-43c0a6b40cf0" containerName="probe" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.521648 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="4bdd15be-7524-4447-bab5-43c0a6b40cf0" containerName="cinder-scheduler" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.522541 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.522619 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.524487 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cinder-scheduler-config-data" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.663281 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a824a183-d65e-4b45-8ecf-72bd3ce9063b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a824a183-d65e-4b45-8ecf-72bd3ce9063b\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.663320 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a824a183-d65e-4b45-8ecf-72bd3ce9063b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a824a183-d65e-4b45-8ecf-72bd3ce9063b\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.663395 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a824a183-d65e-4b45-8ecf-72bd3ce9063b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a824a183-d65e-4b45-8ecf-72bd3ce9063b\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.663438 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a824a183-d65e-4b45-8ecf-72bd3ce9063b-scripts\") pod \"cinder-scheduler-0\" (UID: \"a824a183-d65e-4b45-8ecf-72bd3ce9063b\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.663474 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mvvt2\" (UniqueName: \"kubernetes.io/projected/a824a183-d65e-4b45-8ecf-72bd3ce9063b-kube-api-access-mvvt2\") pod \"cinder-scheduler-0\" (UID: \"a824a183-d65e-4b45-8ecf-72bd3ce9063b\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.663504 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a824a183-d65e-4b45-8ecf-72bd3ce9063b-config-data\") pod \"cinder-scheduler-0\" (UID: \"a824a183-d65e-4b45-8ecf-72bd3ce9063b\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.765404 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a824a183-d65e-4b45-8ecf-72bd3ce9063b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a824a183-d65e-4b45-8ecf-72bd3ce9063b\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.765445 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a824a183-d65e-4b45-8ecf-72bd3ce9063b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a824a183-d65e-4b45-8ecf-72bd3ce9063b\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.765521 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a824a183-d65e-4b45-8ecf-72bd3ce9063b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a824a183-d65e-4b45-8ecf-72bd3ce9063b\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.765532 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-machine-id\" (UniqueName: \"kubernetes.io/host-path/a824a183-d65e-4b45-8ecf-72bd3ce9063b-etc-machine-id\") pod \"cinder-scheduler-0\" (UID: \"a824a183-d65e-4b45-8ecf-72bd3ce9063b\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.765568 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a824a183-d65e-4b45-8ecf-72bd3ce9063b-scripts\") pod \"cinder-scheduler-0\" (UID: \"a824a183-d65e-4b45-8ecf-72bd3ce9063b\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.765607 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mvvt2\" (UniqueName: \"kubernetes.io/projected/a824a183-d65e-4b45-8ecf-72bd3ce9063b-kube-api-access-mvvt2\") pod \"cinder-scheduler-0\" (UID: \"a824a183-d65e-4b45-8ecf-72bd3ce9063b\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.765639 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a824a183-d65e-4b45-8ecf-72bd3ce9063b-config-data\") pod \"cinder-scheduler-0\" (UID: \"a824a183-d65e-4b45-8ecf-72bd3ce9063b\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.778342 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/a824a183-d65e-4b45-8ecf-72bd3ce9063b-config-data-custom\") pod \"cinder-scheduler-0\" (UID: \"a824a183-d65e-4b45-8ecf-72bd3ce9063b\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.778541 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a824a183-d65e-4b45-8ecf-72bd3ce9063b-combined-ca-bundle\") pod \"cinder-scheduler-0\" (UID: \"a824a183-d65e-4b45-8ecf-72bd3ce9063b\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.778688 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/a824a183-d65e-4b45-8ecf-72bd3ce9063b-scripts\") pod \"cinder-scheduler-0\" (UID: \"a824a183-d65e-4b45-8ecf-72bd3ce9063b\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.779091 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a824a183-d65e-4b45-8ecf-72bd3ce9063b-config-data\") pod \"cinder-scheduler-0\" (UID: \"a824a183-d65e-4b45-8ecf-72bd3ce9063b\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.781371 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mvvt2\" (UniqueName: \"kubernetes.io/projected/a824a183-d65e-4b45-8ecf-72bd3ce9063b-kube-api-access-mvvt2\") pod \"cinder-scheduler-0\" (UID: \"a824a183-d65e-4b45-8ecf-72bd3ce9063b\") " pod="openstack/cinder-scheduler-0" Oct 11 04:08:55 crc kubenswrapper[4967]: I1011 04:08:55.836246 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/cinder-scheduler-0" Oct 11 04:08:56 crc kubenswrapper[4967]: I1011 04:08:56.310761 4967 generic.go:334] "Generic (PLEG): container finished" podID="5c82b179-3510-4fb2-9352-d41e7193d221" containerID="08d7b35bd810dabe97b43264d89fe7cfe49649ec8d7c9897c002dfd2a784d9a5" exitCode=1 Oct 11 04:08:56 crc kubenswrapper[4967]: I1011 04:08:56.310901 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-5b5f75f756-jlhwg" event={"ID":"5c82b179-3510-4fb2-9352-d41e7193d221","Type":"ContainerDied","Data":"08d7b35bd810dabe97b43264d89fe7cfe49649ec8d7c9897c002dfd2a784d9a5"} Oct 11 04:08:56 crc kubenswrapper[4967]: I1011 04:08:56.311195 4967 scope.go:117] "RemoveContainer" containerID="5ab54b7897056a999ed147d085d6fa895270af2bd90cb31aa6dd4ff5bcf5712b" Oct 11 04:08:56 crc kubenswrapper[4967]: I1011 04:08:56.311568 4967 scope.go:117] "RemoveContainer" containerID="08d7b35bd810dabe97b43264d89fe7cfe49649ec8d7c9897c002dfd2a784d9a5" Oct 11 04:08:56 crc kubenswrapper[4967]: E1011 04:08:56.311832 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ironic-api pod=ironic-5b5f75f756-jlhwg_openstack(5c82b179-3510-4fb2-9352-d41e7193d221)\"" pod="openstack/ironic-5b5f75f756-jlhwg" podUID="5c82b179-3510-4fb2-9352-d41e7193d221" Oct 11 04:08:56 crc kubenswrapper[4967]: I1011 04:08:56.336839 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eebc2018-61d9-4fe8-ace1-fcb6d817fe80","Type":"ContainerStarted","Data":"c2013cb63bc9eb909af03fc2c4e288a5a06a28fd0bff917277631dba4221e1e2"} Oct 11 04:08:56 crc kubenswrapper[4967]: I1011 04:08:56.354311 4967 generic.go:334] "Generic (PLEG): container finished" podID="706a95a7-728b-40e8-8c4e-221cc836295f" containerID="82f3a6bd6722ea37f904a5fbd419d21d038cfd670bba6e8fa7206aed09197121" exitCode=0 Oct 11 04:08:56 crc kubenswrapper[4967]: I1011 04:08:56.354362 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-zb7zk" event={"ID":"706a95a7-728b-40e8-8c4e-221cc836295f","Type":"ContainerDied","Data":"82f3a6bd6722ea37f904a5fbd419d21d038cfd670bba6e8fa7206aed09197121"} Oct 11 04:08:56 crc kubenswrapper[4967]: I1011 04:08:56.356357 4967 generic.go:334] "Generic (PLEG): container finished" podID="098dc82c-7d36-48e4-bc48-28e8721f6617" containerID="383dd9d4109b281e5b663d9eeb8e34c15b3a8486d8b06d354d56d1399811c062" exitCode=0 Oct 11 04:08:56 crc kubenswrapper[4967]: I1011 04:08:56.356396 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-zth24" event={"ID":"098dc82c-7d36-48e4-bc48-28e8721f6617","Type":"ContainerDied","Data":"383dd9d4109b281e5b663d9eeb8e34c15b3a8486d8b06d354d56d1399811c062"} Oct 11 04:08:56 crc kubenswrapper[4967]: I1011 04:08:56.358136 4967 generic.go:334] "Generic (PLEG): container finished" podID="de9db0f0-79dd-4edc-8022-c32a3a1340a1" containerID="f8b2adf251e6a745e947a0a7231014ac44446c459e8fa03fa0dc9d372d3475dc" exitCode=0 Oct 11 04:08:56 crc kubenswrapper[4967]: I1011 04:08:56.358180 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-ba91-account-create-595xr" event={"ID":"de9db0f0-79dd-4edc-8022-c32a3a1340a1","Type":"ContainerDied","Data":"f8b2adf251e6a745e947a0a7231014ac44446c459e8fa03fa0dc9d372d3475dc"} Oct 11 04:08:56 crc kubenswrapper[4967]: I1011 04:08:56.372591 4967 generic.go:334] "Generic (PLEG): container finished" podID="540b1dfa-8e24-4aed-a3d9-6c2a1e367155" containerID="e018e7db55ec5edeecce65b93021dd2c61dd64dd6b73dd0f4ea82c4c455246a7" exitCode=0 Oct 11 04:08:56 crc kubenswrapper[4967]: I1011 04:08:56.373503 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-zt7cl" event={"ID":"540b1dfa-8e24-4aed-a3d9-6c2a1e367155","Type":"ContainerDied","Data":"e018e7db55ec5edeecce65b93021dd2c61dd64dd6b73dd0f4ea82c4c455246a7"} Oct 11 04:08:56 crc kubenswrapper[4967]: I1011 04:08:56.562303 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/cinder-scheduler-0"] Oct 11 04:08:56 crc kubenswrapper[4967]: I1011 04:08:56.768417 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-zt7cl" Oct 11 04:08:56 crc kubenswrapper[4967]: I1011 04:08:56.825205 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bdd15be-7524-4447-bab5-43c0a6b40cf0" path="/var/lib/kubelet/pods/4bdd15be-7524-4447-bab5-43c0a6b40cf0/volumes" Oct 11 04:08:56 crc kubenswrapper[4967]: I1011 04:08:56.895633 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dtsq\" (UniqueName: \"kubernetes.io/projected/540b1dfa-8e24-4aed-a3d9-6c2a1e367155-kube-api-access-6dtsq\") pod \"540b1dfa-8e24-4aed-a3d9-6c2a1e367155\" (UID: \"540b1dfa-8e24-4aed-a3d9-6c2a1e367155\") " Oct 11 04:08:56 crc kubenswrapper[4967]: I1011 04:08:56.901378 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/540b1dfa-8e24-4aed-a3d9-6c2a1e367155-kube-api-access-6dtsq" (OuterVolumeSpecName: "kube-api-access-6dtsq") pod "540b1dfa-8e24-4aed-a3d9-6c2a1e367155" (UID: "540b1dfa-8e24-4aed-a3d9-6c2a1e367155"). InnerVolumeSpecName "kube-api-access-6dtsq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:08:56 crc kubenswrapper[4967]: I1011 04:08:56.997970 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dtsq\" (UniqueName: \"kubernetes.io/projected/540b1dfa-8e24-4aed-a3d9-6c2a1e367155-kube-api-access-6dtsq\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:57 crc kubenswrapper[4967]: I1011 04:08:57.405453 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a824a183-d65e-4b45-8ecf-72bd3ce9063b","Type":"ContainerStarted","Data":"b47640d2f6cc8fd46af42467e7bd591f5176205e30d07d96722ecaa9e78a2077"} Oct 11 04:08:57 crc kubenswrapper[4967]: I1011 04:08:57.405739 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a824a183-d65e-4b45-8ecf-72bd3ce9063b","Type":"ContainerStarted","Data":"32d0a9e0b88c3536a4db7ea2138670402ae5b1b4363f0b57c903b1bc9fbc953d"} Oct 11 04:08:57 crc kubenswrapper[4967]: I1011 04:08:57.434386 4967 generic.go:334] "Generic (PLEG): container finished" podID="f3fcbf92-aff2-444c-abf1-b23c45ef96e9" containerID="22aca31c65aece17da7d51c55636940bceadde7fd9c8f3b5c5066057250f4000" exitCode=0 Oct 11 04:08:57 crc kubenswrapper[4967]: I1011 04:08:57.434474 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f3fcbf92-aff2-444c-abf1-b23c45ef96e9","Type":"ContainerDied","Data":"22aca31c65aece17da7d51c55636940bceadde7fd9c8f3b5c5066057250f4000"} Oct 11 04:08:57 crc kubenswrapper[4967]: I1011 04:08:57.452432 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-db-create-zt7cl" Oct 11 04:08:57 crc kubenswrapper[4967]: I1011 04:08:57.452425 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-db-create-zt7cl" event={"ID":"540b1dfa-8e24-4aed-a3d9-6c2a1e367155","Type":"ContainerDied","Data":"7288b0044e5bb280ff8bd2888b6619cf178d21b0daa876a455953cc24ae11355"} Oct 11 04:08:57 crc kubenswrapper[4967]: I1011 04:08:57.453119 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7288b0044e5bb280ff8bd2888b6619cf178d21b0daa876a455953cc24ae11355" Oct 11 04:08:57 crc kubenswrapper[4967]: I1011 04:08:57.455656 4967 scope.go:117] "RemoveContainer" containerID="08d7b35bd810dabe97b43264d89fe7cfe49649ec8d7c9897c002dfd2a784d9a5" Oct 11 04:08:57 crc kubenswrapper[4967]: E1011 04:08:57.455971 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-api\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ironic-api pod=ironic-5b5f75f756-jlhwg_openstack(5c82b179-3510-4fb2-9352-d41e7193d221)\"" pod="openstack/ironic-5b5f75f756-jlhwg" podUID="5c82b179-3510-4fb2-9352-d41e7193d221" Oct 11 04:08:57 crc kubenswrapper[4967]: I1011 04:08:57.467326 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eebc2018-61d9-4fe8-ace1-fcb6d817fe80","Type":"ContainerStarted","Data":"49b331cfb61ba6df00a1aebb22e762b74800f8075ba3aef3dc65388ffee26137"} Oct 11 04:08:57 crc kubenswrapper[4967]: I1011 04:08:57.467571 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eebc2018-61d9-4fe8-ace1-fcb6d817fe80","Type":"ContainerStarted","Data":"29a1b90f5cd75564b573dc3d82834afff7c4d05f0f9665c1a7a693e17da9c687"} Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.002404 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-zb7zk" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.150290 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mrqsh\" (UniqueName: \"kubernetes.io/projected/706a95a7-728b-40e8-8c4e-221cc836295f-kube-api-access-mrqsh\") pod \"706a95a7-728b-40e8-8c4e-221cc836295f\" (UID: \"706a95a7-728b-40e8-8c4e-221cc836295f\") " Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.181973 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/706a95a7-728b-40e8-8c4e-221cc836295f-kube-api-access-mrqsh" (OuterVolumeSpecName: "kube-api-access-mrqsh") pod "706a95a7-728b-40e8-8c4e-221cc836295f" (UID: "706a95a7-728b-40e8-8c4e-221cc836295f"). InnerVolumeSpecName "kube-api-access-mrqsh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.253196 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mrqsh\" (UniqueName: \"kubernetes.io/projected/706a95a7-728b-40e8-8c4e-221cc836295f-kube-api-access-mrqsh\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.435091 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.442634 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-zth24" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.470535 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-ba91-account-create-595xr" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.549436 4967 generic.go:334] "Generic (PLEG): container finished" podID="4390a950-7185-4fc0-b152-b5927940032b" containerID="b8c1b1f153783c154d8166fe2c821e9aac2b99008b9620c42dc8206811d3e2aa" exitCode=1 Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.549504 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" event={"ID":"4390a950-7185-4fc0-b152-b5927940032b","Type":"ContainerDied","Data":"b8c1b1f153783c154d8166fe2c821e9aac2b99008b9620c42dc8206811d3e2aa"} Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.549537 4967 scope.go:117] "RemoveContainer" containerID="35a87df2ce5c675c15d00b9d84c1bc636b767b23e69655ff02543264611901bd" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.550160 4967 scope.go:117] "RemoveContainer" containerID="b8c1b1f153783c154d8166fe2c821e9aac2b99008b9620c42dc8206811d3e2aa" Oct 11 04:08:58 crc kubenswrapper[4967]: E1011 04:08:58.550546 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-neutron-agent\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ironic-neutron-agent pod=ironic-neutron-agent-657f6548d9-9fb8w_openstack(4390a950-7185-4fc0-b152-b5927940032b)\"" pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" podUID="4390a950-7185-4fc0-b152-b5927940032b" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.564763 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gmqzl\" (UniqueName: \"kubernetes.io/projected/de9db0f0-79dd-4edc-8022-c32a3a1340a1-kube-api-access-gmqzl\") pod \"de9db0f0-79dd-4edc-8022-c32a3a1340a1\" (UID: \"de9db0f0-79dd-4edc-8022-c32a3a1340a1\") " Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.564820 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-logs\") pod \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") " Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.564878 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4srh\" (UniqueName: \"kubernetes.io/projected/098dc82c-7d36-48e4-bc48-28e8721f6617-kube-api-access-x4srh\") pod \"098dc82c-7d36-48e4-bc48-28e8721f6617\" (UID: \"098dc82c-7d36-48e4-bc48-28e8721f6617\") " Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.564942 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-combined-ca-bundle\") pod \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") " Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.565018 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-config-data\") pod \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") " Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.565050 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-scripts\") pod \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") " Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.565096 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") " Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.565126 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-public-tls-certs\") pod \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") " Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.565159 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hgz4k\" (UniqueName: \"kubernetes.io/projected/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-kube-api-access-hgz4k\") pod \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") " Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.565182 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-httpd-run\") pod \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\" (UID: \"f3fcbf92-aff2-444c-abf1-b23c45ef96e9\") " Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.565708 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "f3fcbf92-aff2-444c-abf1-b23c45ef96e9" (UID: "f3fcbf92-aff2-444c-abf1-b23c45ef96e9"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.566049 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-logs" (OuterVolumeSpecName: "logs") pod "f3fcbf92-aff2-444c-abf1-b23c45ef96e9" (UID: "f3fcbf92-aff2-444c-abf1-b23c45ef96e9"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.582607 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-db-create-zth24" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.582636 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.582696 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-db-create-zth24" event={"ID":"098dc82c-7d36-48e4-bc48-28e8721f6617","Type":"ContainerDied","Data":"9ce5713be9ef6272f759025c19b4643dddaad0710b4a2482d5b2c879096540b6"} Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.582733 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9ce5713be9ef6272f759025c19b4643dddaad0710b4a2482d5b2c879096540b6" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.585518 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/de9db0f0-79dd-4edc-8022-c32a3a1340a1-kube-api-access-gmqzl" (OuterVolumeSpecName: "kube-api-access-gmqzl") pod "de9db0f0-79dd-4edc-8022-c32a3a1340a1" (UID: "de9db0f0-79dd-4edc-8022-c32a3a1340a1"). InnerVolumeSpecName "kube-api-access-gmqzl". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.587973 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-kube-api-access-hgz4k" (OuterVolumeSpecName: "kube-api-access-hgz4k") pod "f3fcbf92-aff2-444c-abf1-b23c45ef96e9" (UID: "f3fcbf92-aff2-444c-abf1-b23c45ef96e9"). InnerVolumeSpecName "kube-api-access-hgz4k". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.588454 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/098dc82c-7d36-48e4-bc48-28e8721f6617-kube-api-access-x4srh" (OuterVolumeSpecName: "kube-api-access-x4srh") pod "098dc82c-7d36-48e4-bc48-28e8721f6617" (UID: "098dc82c-7d36-48e4-bc48-28e8721f6617"). InnerVolumeSpecName "kube-api-access-x4srh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.589788 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"f3fcbf92-aff2-444c-abf1-b23c45ef96e9","Type":"ContainerDied","Data":"54c57f77efaa0ea0680411e499b7a5aa95c6b8d5fea0a28e30ebfdfc07463f01"} Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.589926 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.590310 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage02-crc" (OuterVolumeSpecName: "glance") pod "f3fcbf92-aff2-444c-abf1-b23c45ef96e9" (UID: "f3fcbf92-aff2-444c-abf1-b23c45ef96e9"). InnerVolumeSpecName "local-storage02-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.595202 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-scripts" (OuterVolumeSpecName: "scripts") pod "f3fcbf92-aff2-444c-abf1-b23c45ef96e9" (UID: "f3fcbf92-aff2-444c-abf1-b23c45ef96e9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.603855 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-ba91-account-create-595xr" event={"ID":"de9db0f0-79dd-4edc-8022-c32a3a1340a1","Type":"ContainerDied","Data":"0bacacd3eff1584a6ee5f64df2350f69a6814867989d5ac49c84a39594ef9e41"} Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.603895 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0bacacd3eff1584a6ee5f64df2350f69a6814867989d5ac49c84a39594ef9e41" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.603952 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-ba91-account-create-595xr" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.633581 4967 generic.go:334] "Generic (PLEG): container finished" podID="d8df67d1-7a02-4e18-932a-433412376643" containerID="78ddb9d89814b4e9e287119e168712ef35d876c9d62bc9a8ead0ec4db0ae1f75" exitCode=0 Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.633671 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"d8df67d1-7a02-4e18-932a-433412376643","Type":"ContainerDied","Data":"78ddb9d89814b4e9e287119e168712ef35d876c9d62bc9a8ead0ec4db0ae1f75"} Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.633763 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.655269 4967 generic.go:334] "Generic (PLEG): container finished" podID="92205841-5c2e-4fc7-9328-8a164f5fb8cf" containerID="97c51df86213c4cd58189937091722374bf529014675fdfdb2cbbc6d2816f6dd" exitCode=0 Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.655344 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6554fc45b4-x7btb" event={"ID":"92205841-5c2e-4fc7-9328-8a164f5fb8cf","Type":"ContainerDied","Data":"97c51df86213c4cd58189937091722374bf529014675fdfdb2cbbc6d2816f6dd"} Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.668838 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-db-create-zb7zk" event={"ID":"706a95a7-728b-40e8-8c4e-221cc836295f","Type":"ContainerDied","Data":"6c24146f4bba5945fb8c7ddc6ec6865fc3f1f5b76ad9a0c577864c549c6026cb"} Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.668878 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6c24146f4bba5945fb8c7ddc6ec6865fc3f1f5b76ad9a0c577864c549c6026cb" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.668939 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-db-create-zb7zk" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.669228 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8df67d1-7a02-4e18-932a-433412376643-combined-ca-bundle\") pod \"d8df67d1-7a02-4e18-932a-433412376643\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") " Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.669271 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"glance\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"d8df67d1-7a02-4e18-932a-433412376643\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") " Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.669301 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8df67d1-7a02-4e18-932a-433412376643-config-data\") pod \"d8df67d1-7a02-4e18-932a-433412376643\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") " Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.669393 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d8df67d1-7a02-4e18-932a-433412376643-httpd-run\") pod \"d8df67d1-7a02-4e18-932a-433412376643\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") " Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.669441 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8df67d1-7a02-4e18-932a-433412376643-internal-tls-certs\") pod \"d8df67d1-7a02-4e18-932a-433412376643\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") " Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.669517 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m7jb4\" (UniqueName: \"kubernetes.io/projected/d8df67d1-7a02-4e18-932a-433412376643-kube-api-access-m7jb4\") pod \"d8df67d1-7a02-4e18-932a-433412376643\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") " Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.669563 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8df67d1-7a02-4e18-932a-433412376643-scripts\") pod \"d8df67d1-7a02-4e18-932a-433412376643\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") " Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.669655 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8df67d1-7a02-4e18-932a-433412376643-logs\") pod \"d8df67d1-7a02-4e18-932a-433412376643\" (UID: \"d8df67d1-7a02-4e18-932a-433412376643\") " Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.680237 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage04-crc" (OuterVolumeSpecName: "glance") pod "d8df67d1-7a02-4e18-932a-433412376643" (UID: "d8df67d1-7a02-4e18-932a-433412376643"). InnerVolumeSpecName "local-storage04-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.681057 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8df67d1-7a02-4e18-932a-433412376643-httpd-run" (OuterVolumeSpecName: "httpd-run") pod "d8df67d1-7a02-4e18-932a-433412376643" (UID: "d8df67d1-7a02-4e18-932a-433412376643"). InnerVolumeSpecName "httpd-run". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.681416 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d8df67d1-7a02-4e18-932a-433412376643-logs" (OuterVolumeSpecName: "logs") pod "d8df67d1-7a02-4e18-932a-433412376643" (UID: "d8df67d1-7a02-4e18-932a-433412376643"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.686437 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4srh\" (UniqueName: \"kubernetes.io/projected/098dc82c-7d36-48e4-bc48-28e8721f6617-kube-api-access-x4srh\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.686468 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.686487 4967 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" " Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.686497 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hgz4k\" (UniqueName: \"kubernetes.io/projected/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-kube-api-access-hgz4k\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.686507 4967 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.686519 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gmqzl\" (UniqueName: \"kubernetes.io/projected/de9db0f0-79dd-4edc-8022-c32a3a1340a1-kube-api-access-gmqzl\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.686530 4967 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-logs\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.693539 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d8df67d1-7a02-4e18-932a-433412376643-kube-api-access-m7jb4" (OuterVolumeSpecName: "kube-api-access-m7jb4") pod "d8df67d1-7a02-4e18-932a-433412376643" (UID: "d8df67d1-7a02-4e18-932a-433412376643"). InnerVolumeSpecName "kube-api-access-m7jb4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.733315 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "f3fcbf92-aff2-444c-abf1-b23c45ef96e9" (UID: "f3fcbf92-aff2-444c-abf1-b23c45ef96e9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.737459 4967 scope.go:117] "RemoveContainer" containerID="22aca31c65aece17da7d51c55636940bceadde7fd9c8f3b5c5066057250f4000" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.749195 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8df67d1-7a02-4e18-932a-433412376643-scripts" (OuterVolumeSpecName: "scripts") pod "d8df67d1-7a02-4e18-932a-433412376643" (UID: "d8df67d1-7a02-4e18-932a-433412376643"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.773144 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "f3fcbf92-aff2-444c-abf1-b23c45ef96e9" (UID: "f3fcbf92-aff2-444c-abf1-b23c45ef96e9"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.788478 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m7jb4\" (UniqueName: \"kubernetes.io/projected/d8df67d1-7a02-4e18-932a-433412376643-kube-api-access-m7jb4\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.788500 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d8df67d1-7a02-4e18-932a-433412376643-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.788509 4967 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.788517 4967 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/d8df67d1-7a02-4e18-932a-433412376643-logs\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.788536 4967 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" " Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.788546 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.788555 4967 reconciler_common.go:293] "Volume detached for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/d8df67d1-7a02-4e18-932a-433412376643-httpd-run\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.788886 4967 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage02-crc" (UniqueName: "kubernetes.io/local-volume/local-storage02-crc") on node "crc" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.791623 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-config-data" (OuterVolumeSpecName: "config-data") pod "f3fcbf92-aff2-444c-abf1-b23c45ef96e9" (UID: "f3fcbf92-aff2-444c-abf1-b23c45ef96e9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.824369 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8df67d1-7a02-4e18-932a-433412376643-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d8df67d1-7a02-4e18-932a-433412376643" (UID: "d8df67d1-7a02-4e18-932a-433412376643"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.825754 4967 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage04-crc" (UniqueName: "kubernetes.io/local-volume/local-storage04-crc") on node "crc" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.842451 4967 scope.go:117] "RemoveContainer" containerID="0e68c39cb6ce4b2246948b7da0836b7edf0d64a7d6f59868bf31e93d2857f817" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.845570 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8df67d1-7a02-4e18-932a-433412376643-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "d8df67d1-7a02-4e18-932a-433412376643" (UID: "d8df67d1-7a02-4e18-932a-433412376643"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.864166 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d8df67d1-7a02-4e18-932a-433412376643-config-data" (OuterVolumeSpecName: "config-data") pod "d8df67d1-7a02-4e18-932a-433412376643" (UID: "d8df67d1-7a02-4e18-932a-433412376643"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.900835 4967 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/d8df67d1-7a02-4e18-932a-433412376643-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.900871 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/f3fcbf92-aff2-444c-abf1-b23c45ef96e9-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.900881 4967 reconciler_common.go:293] "Volume detached for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.900892 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d8df67d1-7a02-4e18-932a-433412376643-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.900926 4967 reconciler_common.go:293] "Volume detached for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.900936 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d8df67d1-7a02-4e18-932a-433412376643-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.943386 4967 scope.go:117] "RemoveContainer" containerID="78ddb9d89814b4e9e287119e168712ef35d876c9d62bc9a8ead0ec4db0ae1f75" Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.948501 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 11 04:08:58 crc kubenswrapper[4967]: I1011 04:08:58.979565 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.012617 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-external-api-0"] Oct 11 04:08:59 crc kubenswrapper[4967]: E1011 04:08:59.012986 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="098dc82c-7d36-48e4-bc48-28e8721f6617" containerName="mariadb-database-create" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.013003 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="098dc82c-7d36-48e4-bc48-28e8721f6617" containerName="mariadb-database-create" Oct 11 04:08:59 crc kubenswrapper[4967]: E1011 04:08:59.013012 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="706a95a7-728b-40e8-8c4e-221cc836295f" containerName="mariadb-database-create" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.013018 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="706a95a7-728b-40e8-8c4e-221cc836295f" containerName="mariadb-database-create" Oct 11 04:08:59 crc kubenswrapper[4967]: E1011 04:08:59.013025 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8df67d1-7a02-4e18-932a-433412376643" containerName="glance-httpd" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.013031 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8df67d1-7a02-4e18-932a-433412376643" containerName="glance-httpd" Oct 11 04:08:59 crc kubenswrapper[4967]: E1011 04:08:59.013040 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3fcbf92-aff2-444c-abf1-b23c45ef96e9" containerName="glance-httpd" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.013045 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3fcbf92-aff2-444c-abf1-b23c45ef96e9" containerName="glance-httpd" Oct 11 04:08:59 crc kubenswrapper[4967]: E1011 04:08:59.013055 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="540b1dfa-8e24-4aed-a3d9-6c2a1e367155" containerName="mariadb-database-create" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.013062 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="540b1dfa-8e24-4aed-a3d9-6c2a1e367155" containerName="mariadb-database-create" Oct 11 04:08:59 crc kubenswrapper[4967]: E1011 04:08:59.013091 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f3fcbf92-aff2-444c-abf1-b23c45ef96e9" containerName="glance-log" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.013097 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="f3fcbf92-aff2-444c-abf1-b23c45ef96e9" containerName="glance-log" Oct 11 04:08:59 crc kubenswrapper[4967]: E1011 04:08:59.013108 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d8df67d1-7a02-4e18-932a-433412376643" containerName="glance-log" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.013114 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="d8df67d1-7a02-4e18-932a-433412376643" containerName="glance-log" Oct 11 04:08:59 crc kubenswrapper[4967]: E1011 04:08:59.013142 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="de9db0f0-79dd-4edc-8022-c32a3a1340a1" containerName="mariadb-account-create" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.013148 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="de9db0f0-79dd-4edc-8022-c32a3a1340a1" containerName="mariadb-account-create" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.013313 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="de9db0f0-79dd-4edc-8022-c32a3a1340a1" containerName="mariadb-account-create" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.013326 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="540b1dfa-8e24-4aed-a3d9-6c2a1e367155" containerName="mariadb-database-create" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.013341 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3fcbf92-aff2-444c-abf1-b23c45ef96e9" containerName="glance-httpd" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.013351 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="f3fcbf92-aff2-444c-abf1-b23c45ef96e9" containerName="glance-log" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.013358 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8df67d1-7a02-4e18-932a-433412376643" containerName="glance-httpd" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.013367 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="098dc82c-7d36-48e4-bc48-28e8721f6617" containerName="mariadb-database-create" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.013378 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="706a95a7-728b-40e8-8c4e-221cc836295f" containerName="mariadb-database-create" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.013389 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="d8df67d1-7a02-4e18-932a-433412376643" containerName="glance-log" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.014298 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.017236 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-external-config-data" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.017533 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-public-svc" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.017745 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-glance-dockercfg-2d8h2" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.018053 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-scripts" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.021334 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6554fc45b4-x7btb" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.039412 4967 scope.go:117] "RemoveContainer" containerID="223177e0c8f9782ee3f095403a0314a51431ed701e59885aa682f2114dd478b2" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.040037 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.078118 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.091844 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.100177 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 11 04:08:59 crc kubenswrapper[4967]: E1011 04:08:59.100550 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92205841-5c2e-4fc7-9328-8a164f5fb8cf" containerName="neutron-api" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.100561 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="92205841-5c2e-4fc7-9328-8a164f5fb8cf" containerName="neutron-api" Oct 11 04:08:59 crc kubenswrapper[4967]: E1011 04:08:59.100599 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="92205841-5c2e-4fc7-9328-8a164f5fb8cf" containerName="neutron-httpd" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.100605 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="92205841-5c2e-4fc7-9328-8a164f5fb8cf" containerName="neutron-httpd" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.100790 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="92205841-5c2e-4fc7-9328-8a164f5fb8cf" containerName="neutron-api" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.100812 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="92205841-5c2e-4fc7-9328-8a164f5fb8cf" containerName="neutron-httpd" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.101961 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.104088 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/92205841-5c2e-4fc7-9328-8a164f5fb8cf-config\") pod \"92205841-5c2e-4fc7-9328-8a164f5fb8cf\" (UID: \"92205841-5c2e-4fc7-9328-8a164f5fb8cf\") " Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.104361 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/92205841-5c2e-4fc7-9328-8a164f5fb8cf-ovndb-tls-certs\") pod \"92205841-5c2e-4fc7-9328-8a164f5fb8cf\" (UID: \"92205841-5c2e-4fc7-9328-8a164f5fb8cf\") " Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.104612 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92205841-5c2e-4fc7-9328-8a164f5fb8cf-combined-ca-bundle\") pod \"92205841-5c2e-4fc7-9328-8a164f5fb8cf\" (UID: \"92205841-5c2e-4fc7-9328-8a164f5fb8cf\") " Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.104679 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/92205841-5c2e-4fc7-9328-8a164f5fb8cf-httpd-config\") pod \"92205841-5c2e-4fc7-9328-8a164f5fb8cf\" (UID: \"92205841-5c2e-4fc7-9328-8a164f5fb8cf\") " Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.104705 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ld8gb\" (UniqueName: \"kubernetes.io/projected/92205841-5c2e-4fc7-9328-8a164f5fb8cf-kube-api-access-ld8gb\") pod \"92205841-5c2e-4fc7-9328-8a164f5fb8cf\" (UID: \"92205841-5c2e-4fc7-9328-8a164f5fb8cf\") " Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.104981 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"662b8e47-db97-4003-892a-48c25cba4589\") " pod="openstack/glance-default-external-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.105004 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/662b8e47-db97-4003-892a-48c25cba4589-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"662b8e47-db97-4003-892a-48c25cba4589\") " pod="openstack/glance-default-external-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.105085 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mwdkf\" (UniqueName: \"kubernetes.io/projected/662b8e47-db97-4003-892a-48c25cba4589-kube-api-access-mwdkf\") pod \"glance-default-external-api-0\" (UID: \"662b8e47-db97-4003-892a-48c25cba4589\") " pod="openstack/glance-default-external-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.105107 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/662b8e47-db97-4003-892a-48c25cba4589-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"662b8e47-db97-4003-892a-48c25cba4589\") " pod="openstack/glance-default-external-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.105128 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/662b8e47-db97-4003-892a-48c25cba4589-config-data\") pod \"glance-default-external-api-0\" (UID: \"662b8e47-db97-4003-892a-48c25cba4589\") " pod="openstack/glance-default-external-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.105208 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/662b8e47-db97-4003-892a-48c25cba4589-scripts\") pod \"glance-default-external-api-0\" (UID: \"662b8e47-db97-4003-892a-48c25cba4589\") " pod="openstack/glance-default-external-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.105238 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/662b8e47-db97-4003-892a-48c25cba4589-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"662b8e47-db97-4003-892a-48c25cba4589\") " pod="openstack/glance-default-external-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.106402 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/662b8e47-db97-4003-892a-48c25cba4589-logs\") pod \"glance-default-external-api-0\" (UID: \"662b8e47-db97-4003-892a-48c25cba4589\") " pod="openstack/glance-default-external-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.107571 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"glance-default-internal-config-data" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.108230 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-glance-default-internal-svc" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.140666 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92205841-5c2e-4fc7-9328-8a164f5fb8cf-httpd-config" (OuterVolumeSpecName: "httpd-config") pod "92205841-5c2e-4fc7-9328-8a164f5fb8cf" (UID: "92205841-5c2e-4fc7-9328-8a164f5fb8cf"). InnerVolumeSpecName "httpd-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.140867 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/92205841-5c2e-4fc7-9328-8a164f5fb8cf-kube-api-access-ld8gb" (OuterVolumeSpecName: "kube-api-access-ld8gb") pod "92205841-5c2e-4fc7-9328-8a164f5fb8cf" (UID: "92205841-5c2e-4fc7-9328-8a164f5fb8cf"). InnerVolumeSpecName "kube-api-access-ld8gb". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.189537 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.210426 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"53708cdd-1297-4302-9d8a-e00ba57c01b3\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.211309 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/662b8e47-db97-4003-892a-48c25cba4589-scripts\") pod \"glance-default-external-api-0\" (UID: \"662b8e47-db97-4003-892a-48c25cba4589\") " pod="openstack/glance-default-external-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.211405 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/662b8e47-db97-4003-892a-48c25cba4589-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"662b8e47-db97-4003-892a-48c25cba4589\") " pod="openstack/glance-default-external-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.211432 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/53708cdd-1297-4302-9d8a-e00ba57c01b3-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"53708cdd-1297-4302-9d8a-e00ba57c01b3\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.211506 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/662b8e47-db97-4003-892a-48c25cba4589-logs\") pod \"glance-default-external-api-0\" (UID: \"662b8e47-db97-4003-892a-48c25cba4589\") " pod="openstack/glance-default-external-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.211535 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53708cdd-1297-4302-9d8a-e00ba57c01b3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"53708cdd-1297-4302-9d8a-e00ba57c01b3\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.211583 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53708cdd-1297-4302-9d8a-e00ba57c01b3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"53708cdd-1297-4302-9d8a-e00ba57c01b3\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.211613 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"662b8e47-db97-4003-892a-48c25cba4589\") " pod="openstack/glance-default-external-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.211630 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53708cdd-1297-4302-9d8a-e00ba57c01b3-logs\") pod \"glance-default-internal-api-0\" (UID: \"53708cdd-1297-4302-9d8a-e00ba57c01b3\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.211665 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/662b8e47-db97-4003-892a-48c25cba4589-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"662b8e47-db97-4003-892a-48c25cba4589\") " pod="openstack/glance-default-external-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.211723 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/53708cdd-1297-4302-9d8a-e00ba57c01b3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"53708cdd-1297-4302-9d8a-e00ba57c01b3\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.211783 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mwdkf\" (UniqueName: \"kubernetes.io/projected/662b8e47-db97-4003-892a-48c25cba4589-kube-api-access-mwdkf\") pod \"glance-default-external-api-0\" (UID: \"662b8e47-db97-4003-892a-48c25cba4589\") " pod="openstack/glance-default-external-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.211817 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/662b8e47-db97-4003-892a-48c25cba4589-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"662b8e47-db97-4003-892a-48c25cba4589\") " pod="openstack/glance-default-external-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.211849 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53708cdd-1297-4302-9d8a-e00ba57c01b3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"53708cdd-1297-4302-9d8a-e00ba57c01b3\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.211869 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/662b8e47-db97-4003-892a-48c25cba4589-config-data\") pod \"glance-default-external-api-0\" (UID: \"662b8e47-db97-4003-892a-48c25cba4589\") " pod="openstack/glance-default-external-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.212784 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ct22p\" (UniqueName: \"kubernetes.io/projected/53708cdd-1297-4302-9d8a-e00ba57c01b3-kube-api-access-ct22p\") pod \"glance-default-internal-api-0\" (UID: \"53708cdd-1297-4302-9d8a-e00ba57c01b3\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.213360 4967 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"662b8e47-db97-4003-892a-48c25cba4589\") device mount path \"/mnt/openstack/pv02\"" pod="openstack/glance-default-external-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.215466 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/662b8e47-db97-4003-892a-48c25cba4589-logs\") pod \"glance-default-external-api-0\" (UID: \"662b8e47-db97-4003-892a-48c25cba4589\") " pod="openstack/glance-default-external-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.215568 4967 reconciler_common.go:293] "Volume detached for volume \"httpd-config\" (UniqueName: \"kubernetes.io/secret/92205841-5c2e-4fc7-9328-8a164f5fb8cf-httpd-config\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.215597 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ld8gb\" (UniqueName: \"kubernetes.io/projected/92205841-5c2e-4fc7-9328-8a164f5fb8cf-kube-api-access-ld8gb\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.215594 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/662b8e47-db97-4003-892a-48c25cba4589-httpd-run\") pod \"glance-default-external-api-0\" (UID: \"662b8e47-db97-4003-892a-48c25cba4589\") " pod="openstack/glance-default-external-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.221911 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/662b8e47-db97-4003-892a-48c25cba4589-public-tls-certs\") pod \"glance-default-external-api-0\" (UID: \"662b8e47-db97-4003-892a-48c25cba4589\") " pod="openstack/glance-default-external-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.224439 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/662b8e47-db97-4003-892a-48c25cba4589-scripts\") pod \"glance-default-external-api-0\" (UID: \"662b8e47-db97-4003-892a-48c25cba4589\") " pod="openstack/glance-default-external-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.250050 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/662b8e47-db97-4003-892a-48c25cba4589-combined-ca-bundle\") pod \"glance-default-external-api-0\" (UID: \"662b8e47-db97-4003-892a-48c25cba4589\") " pod="openstack/glance-default-external-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.251153 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mwdkf\" (UniqueName: \"kubernetes.io/projected/662b8e47-db97-4003-892a-48c25cba4589-kube-api-access-mwdkf\") pod \"glance-default-external-api-0\" (UID: \"662b8e47-db97-4003-892a-48c25cba4589\") " pod="openstack/glance-default-external-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.255987 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/662b8e47-db97-4003-892a-48c25cba4589-config-data\") pod \"glance-default-external-api-0\" (UID: \"662b8e47-db97-4003-892a-48c25cba4589\") " pod="openstack/glance-default-external-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.283650 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-85b58c7f48-9dgpg" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.323766 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ct22p\" (UniqueName: \"kubernetes.io/projected/53708cdd-1297-4302-9d8a-e00ba57c01b3-kube-api-access-ct22p\") pod \"glance-default-internal-api-0\" (UID: \"53708cdd-1297-4302-9d8a-e00ba57c01b3\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.323860 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"53708cdd-1297-4302-9d8a-e00ba57c01b3\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.323903 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/53708cdd-1297-4302-9d8a-e00ba57c01b3-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"53708cdd-1297-4302-9d8a-e00ba57c01b3\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.323956 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53708cdd-1297-4302-9d8a-e00ba57c01b3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"53708cdd-1297-4302-9d8a-e00ba57c01b3\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.323991 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53708cdd-1297-4302-9d8a-e00ba57c01b3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"53708cdd-1297-4302-9d8a-e00ba57c01b3\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.324018 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53708cdd-1297-4302-9d8a-e00ba57c01b3-logs\") pod \"glance-default-internal-api-0\" (UID: \"53708cdd-1297-4302-9d8a-e00ba57c01b3\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.324053 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/53708cdd-1297-4302-9d8a-e00ba57c01b3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"53708cdd-1297-4302-9d8a-e00ba57c01b3\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.324111 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53708cdd-1297-4302-9d8a-e00ba57c01b3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"53708cdd-1297-4302-9d8a-e00ba57c01b3\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.324788 4967 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"53708cdd-1297-4302-9d8a-e00ba57c01b3\") device mount path \"/mnt/openstack/pv04\"" pod="openstack/glance-default-internal-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.325130 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"httpd-run\" (UniqueName: \"kubernetes.io/empty-dir/53708cdd-1297-4302-9d8a-e00ba57c01b3-httpd-run\") pod \"glance-default-internal-api-0\" (UID: \"53708cdd-1297-4302-9d8a-e00ba57c01b3\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.325340 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/53708cdd-1297-4302-9d8a-e00ba57c01b3-logs\") pod \"glance-default-internal-api-0\" (UID: \"53708cdd-1297-4302-9d8a-e00ba57c01b3\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.352653 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage02-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage02-crc\") pod \"glance-default-external-api-0\" (UID: \"662b8e47-db97-4003-892a-48c25cba4589\") " pod="openstack/glance-default-external-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.354920 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/53708cdd-1297-4302-9d8a-e00ba57c01b3-internal-tls-certs\") pod \"glance-default-internal-api-0\" (UID: \"53708cdd-1297-4302-9d8a-e00ba57c01b3\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.356712 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ct22p\" (UniqueName: \"kubernetes.io/projected/53708cdd-1297-4302-9d8a-e00ba57c01b3-kube-api-access-ct22p\") pod \"glance-default-internal-api-0\" (UID: \"53708cdd-1297-4302-9d8a-e00ba57c01b3\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.365026 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/53708cdd-1297-4302-9d8a-e00ba57c01b3-combined-ca-bundle\") pod \"glance-default-internal-api-0\" (UID: \"53708cdd-1297-4302-9d8a-e00ba57c01b3\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.368744 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/53708cdd-1297-4302-9d8a-e00ba57c01b3-scripts\") pod \"glance-default-internal-api-0\" (UID: \"53708cdd-1297-4302-9d8a-e00ba57c01b3\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.393164 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-5b5f75f756-jlhwg"] Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.393433 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ironic-5b5f75f756-jlhwg" podUID="5c82b179-3510-4fb2-9352-d41e7193d221" containerName="ironic-api-log" containerID="cri-o://565f03d42b75e92dcace4c1ad5111a96bc6b2d2fbcbc46bc88b7137676ce22f6" gracePeriod=60 Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.402547 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-external-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.425869 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/53708cdd-1297-4302-9d8a-e00ba57c01b3-config-data\") pod \"glance-default-internal-api-0\" (UID: \"53708cdd-1297-4302-9d8a-e00ba57c01b3\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.435609 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92205841-5c2e-4fc7-9328-8a164f5fb8cf-config" (OuterVolumeSpecName: "config") pod "92205841-5c2e-4fc7-9328-8a164f5fb8cf" (UID: "92205841-5c2e-4fc7-9328-8a164f5fb8cf"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.444539 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.464532 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/cinder-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.476875 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-5b5f75f756-jlhwg" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.495466 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92205841-5c2e-4fc7-9328-8a164f5fb8cf-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "92205841-5c2e-4fc7-9328-8a164f5fb8cf" (UID: "92205841-5c2e-4fc7-9328-8a164f5fb8cf"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.527575 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/92205841-5c2e-4fc7-9328-8a164f5fb8cf-config\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.527611 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/92205841-5c2e-4fc7-9328-8a164f5fb8cf-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.544579 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage04-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage04-crc\") pod \"glance-default-internal-api-0\" (UID: \"53708cdd-1297-4302-9d8a-e00ba57c01b3\") " pod="openstack/glance-default-internal-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.640361 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/92205841-5c2e-4fc7-9328-8a164f5fb8cf-ovndb-tls-certs" (OuterVolumeSpecName: "ovndb-tls-certs") pod "92205841-5c2e-4fc7-9328-8a164f5fb8cf" (UID: "92205841-5c2e-4fc7-9328-8a164f5fb8cf"). InnerVolumeSpecName "ovndb-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.670845 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/glance-default-internal-api-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.725682 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/neutron-6554fc45b4-x7btb" event={"ID":"92205841-5c2e-4fc7-9328-8a164f5fb8cf","Type":"ContainerDied","Data":"8689a5db47306a95c7d6f1322eaa82093ce6843fba7dbb3f3f9c41ee4440cbd9"} Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.725749 4967 scope.go:117] "RemoveContainer" containerID="50f50943493197ea90b614621946c4164887b9cfb4c14fbe86b0b763dd2f42b6" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.725905 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/neutron-6554fc45b4-x7btb" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.734009 4967 reconciler_common.go:293] "Volume detached for volume \"ovndb-tls-certs\" (UniqueName: \"kubernetes.io/secret/92205841-5c2e-4fc7-9328-8a164f5fb8cf-ovndb-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.785492 4967 generic.go:334] "Generic (PLEG): container finished" podID="5c82b179-3510-4fb2-9352-d41e7193d221" containerID="565f03d42b75e92dcace4c1ad5111a96bc6b2d2fbcbc46bc88b7137676ce22f6" exitCode=143 Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.785564 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-5b5f75f756-jlhwg" event={"ID":"5c82b179-3510-4fb2-9352-d41e7193d221","Type":"ContainerDied","Data":"565f03d42b75e92dcace4c1ad5111a96bc6b2d2fbcbc46bc88b7137676ce22f6"} Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.794295 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eebc2018-61d9-4fe8-ace1-fcb6d817fe80","Type":"ContainerStarted","Data":"b7b9e2f154cda6c793b1118b11625c7241f0fadc381fc945277f06af7eefdc40"} Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.794453 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eebc2018-61d9-4fe8-ace1-fcb6d817fe80" containerName="ceilometer-central-agent" containerID="cri-o://c2013cb63bc9eb909af03fc2c4e288a5a06a28fd0bff917277631dba4221e1e2" gracePeriod=30 Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.794687 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.794966 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eebc2018-61d9-4fe8-ace1-fcb6d817fe80" containerName="proxy-httpd" containerID="cri-o://b7b9e2f154cda6c793b1118b11625c7241f0fadc381fc945277f06af7eefdc40" gracePeriod=30 Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.795015 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eebc2018-61d9-4fe8-ace1-fcb6d817fe80" containerName="sg-core" containerID="cri-o://49b331cfb61ba6df00a1aebb22e762b74800f8075ba3aef3dc65388ffee26137" gracePeriod=30 Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.795048 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="eebc2018-61d9-4fe8-ace1-fcb6d817fe80" containerName="ceilometer-notification-agent" containerID="cri-o://29a1b90f5cd75564b573dc3d82834afff7c4d05f0f9665c1a7a693e17da9c687" gracePeriod=30 Oct 11 04:08:59 crc kubenswrapper[4967]: E1011 04:08:59.798809 4967 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod5c82b179_3510_4fb2_9352_d41e7193d221.slice/crio-conmon-565f03d42b75e92dcace4c1ad5111a96bc6b2d2fbcbc46bc88b7137676ce22f6.scope\": RecentStats: unable to find data in memory cache]" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.809541 4967 scope.go:117] "RemoveContainer" containerID="b8c1b1f153783c154d8166fe2c821e9aac2b99008b9620c42dc8206811d3e2aa" Oct 11 04:08:59 crc kubenswrapper[4967]: E1011 04:08:59.809898 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-neutron-agent\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ironic-neutron-agent pod=ironic-neutron-agent-657f6548d9-9fb8w_openstack(4390a950-7185-4fc0-b152-b5927940032b)\"" pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" podUID="4390a950-7185-4fc0-b152-b5927940032b" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.830597 4967 scope.go:117] "RemoveContainer" containerID="97c51df86213c4cd58189937091722374bf529014675fdfdb2cbbc6d2816f6dd" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.837154 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/cinder-scheduler-0" event={"ID":"a824a183-d65e-4b45-8ecf-72bd3ce9063b","Type":"ContainerStarted","Data":"d153aaa701ad83e8ef665ab9dd46be7999eff9f1037359fe51867f220e90c6dd"} Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.926394 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/cinder-scheduler-0" podStartSLOduration=4.9263755 podStartE2EDuration="4.9263755s" podCreationTimestamp="2025-10-11 04:08:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:08:59.918380598 +0000 UTC m=+1067.881589551" watchObservedRunningTime="2025-10-11 04:08:59.9263755 +0000 UTC m=+1067.889584433" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.937947 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=4.303233087 podStartE2EDuration="8.937930654s" podCreationTimestamp="2025-10-11 04:08:51 +0000 UTC" firstStartedPulling="2025-10-11 04:08:54.154336089 +0000 UTC m=+1062.117545022" lastFinishedPulling="2025-10-11 04:08:58.789033656 +0000 UTC m=+1066.752242589" observedRunningTime="2025-10-11 04:08:59.8784695 +0000 UTC m=+1067.841678423" watchObservedRunningTime="2025-10-11 04:08:59.937930654 +0000 UTC m=+1067.901139587" Oct 11 04:08:59 crc kubenswrapper[4967]: I1011 04:08:59.983492 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-6554fc45b4-x7btb"] Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.016925 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-6554fc45b4-x7btb"] Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.210110 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-5b5f75f756-jlhwg" Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.346653 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c82b179-3510-4fb2-9352-d41e7193d221-scripts\") pod \"5c82b179-3510-4fb2-9352-d41e7193d221\" (UID: \"5c82b179-3510-4fb2-9352-d41e7193d221\") " Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.347113 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/5c82b179-3510-4fb2-9352-d41e7193d221-etc-podinfo\") pod \"5c82b179-3510-4fb2-9352-d41e7193d221\" (UID: \"5c82b179-3510-4fb2-9352-d41e7193d221\") " Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.347191 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/5c82b179-3510-4fb2-9352-d41e7193d221-config-data-merged\") pod \"5c82b179-3510-4fb2-9352-d41e7193d221\" (UID: \"5c82b179-3510-4fb2-9352-d41e7193d221\") " Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.347228 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c82b179-3510-4fb2-9352-d41e7193d221-logs\") pod \"5c82b179-3510-4fb2-9352-d41e7193d221\" (UID: \"5c82b179-3510-4fb2-9352-d41e7193d221\") " Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.347273 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c82b179-3510-4fb2-9352-d41e7193d221-combined-ca-bundle\") pod \"5c82b179-3510-4fb2-9352-d41e7193d221\" (UID: \"5c82b179-3510-4fb2-9352-d41e7193d221\") " Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.347319 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5c82b179-3510-4fb2-9352-d41e7193d221-config-data-custom\") pod \"5c82b179-3510-4fb2-9352-d41e7193d221\" (UID: \"5c82b179-3510-4fb2-9352-d41e7193d221\") " Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.347341 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c82b179-3510-4fb2-9352-d41e7193d221-config-data\") pod \"5c82b179-3510-4fb2-9352-d41e7193d221\" (UID: \"5c82b179-3510-4fb2-9352-d41e7193d221\") " Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.347406 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hjn7n\" (UniqueName: \"kubernetes.io/projected/5c82b179-3510-4fb2-9352-d41e7193d221-kube-api-access-hjn7n\") pod \"5c82b179-3510-4fb2-9352-d41e7193d221\" (UID: \"5c82b179-3510-4fb2-9352-d41e7193d221\") " Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.350562 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c82b179-3510-4fb2-9352-d41e7193d221-logs" (OuterVolumeSpecName: "logs") pod "5c82b179-3510-4fb2-9352-d41e7193d221" (UID: "5c82b179-3510-4fb2-9352-d41e7193d221"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.353971 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5c82b179-3510-4fb2-9352-d41e7193d221-config-data-merged" (OuterVolumeSpecName: "config-data-merged") pod "5c82b179-3510-4fb2-9352-d41e7193d221" (UID: "5c82b179-3510-4fb2-9352-d41e7193d221"). InnerVolumeSpecName "config-data-merged". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.354286 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5c82b179-3510-4fb2-9352-d41e7193d221-kube-api-access-hjn7n" (OuterVolumeSpecName: "kube-api-access-hjn7n") pod "5c82b179-3510-4fb2-9352-d41e7193d221" (UID: "5c82b179-3510-4fb2-9352-d41e7193d221"). InnerVolumeSpecName "kube-api-access-hjn7n". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.363301 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c82b179-3510-4fb2-9352-d41e7193d221-config-data-custom" (OuterVolumeSpecName: "config-data-custom") pod "5c82b179-3510-4fb2-9352-d41e7193d221" (UID: "5c82b179-3510-4fb2-9352-d41e7193d221"). InnerVolumeSpecName "config-data-custom". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.364887 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c82b179-3510-4fb2-9352-d41e7193d221-scripts" (OuterVolumeSpecName: "scripts") pod "5c82b179-3510-4fb2-9352-d41e7193d221" (UID: "5c82b179-3510-4fb2-9352-d41e7193d221"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.368157 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/5c82b179-3510-4fb2-9352-d41e7193d221-etc-podinfo" (OuterVolumeSpecName: "etc-podinfo") pod "5c82b179-3510-4fb2-9352-d41e7193d221" (UID: "5c82b179-3510-4fb2-9352-d41e7193d221"). InnerVolumeSpecName "etc-podinfo". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.379646 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-external-api-0"] Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.412569 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c82b179-3510-4fb2-9352-d41e7193d221-config-data" (OuterVolumeSpecName: "config-data") pod "5c82b179-3510-4fb2-9352-d41e7193d221" (UID: "5c82b179-3510-4fb2-9352-d41e7193d221"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.418908 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5c82b179-3510-4fb2-9352-d41e7193d221-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "5c82b179-3510-4fb2-9352-d41e7193d221" (UID: "5c82b179-3510-4fb2-9352-d41e7193d221"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:09:00 crc kubenswrapper[4967]: W1011 04:09:00.419588 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod662b8e47_db97_4003_892a_48c25cba4589.slice/crio-873ed5ec64809c56f14048935649e48cc79d4aa6f8ce30c9310e0076831baf73 WatchSource:0}: Error finding container 873ed5ec64809c56f14048935649e48cc79d4aa6f8ce30c9310e0076831baf73: Status 404 returned error can't find the container with id 873ed5ec64809c56f14048935649e48cc79d4aa6f8ce30c9310e0076831baf73 Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.449665 4967 reconciler_common.go:293] "Volume detached for volume \"config-data-custom\" (UniqueName: \"kubernetes.io/secret/5c82b179-3510-4fb2-9352-d41e7193d221-config-data-custom\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.449707 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/5c82b179-3510-4fb2-9352-d41e7193d221-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.449719 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hjn7n\" (UniqueName: \"kubernetes.io/projected/5c82b179-3510-4fb2-9352-d41e7193d221-kube-api-access-hjn7n\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.449729 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/5c82b179-3510-4fb2-9352-d41e7193d221-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.449737 4967 reconciler_common.go:293] "Volume detached for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/5c82b179-3510-4fb2-9352-d41e7193d221-etc-podinfo\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.449745 4967 reconciler_common.go:293] "Volume detached for volume \"config-data-merged\" (UniqueName: \"kubernetes.io/empty-dir/5c82b179-3510-4fb2-9352-d41e7193d221-config-data-merged\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.449783 4967 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/5c82b179-3510-4fb2-9352-d41e7193d221-logs\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.449790 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/5c82b179-3510-4fb2-9352-d41e7193d221-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.592918 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/glance-default-internal-api-0"] Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.825693 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92205841-5c2e-4fc7-9328-8a164f5fb8cf" path="/var/lib/kubelet/pods/92205841-5c2e-4fc7-9328-8a164f5fb8cf/volumes" Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.826619 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d8df67d1-7a02-4e18-932a-433412376643" path="/var/lib/kubelet/pods/d8df67d1-7a02-4e18-932a-433412376643/volumes" Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.827347 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f3fcbf92-aff2-444c-abf1-b23c45ef96e9" path="/var/lib/kubelet/pods/f3fcbf92-aff2-444c-abf1-b23c45ef96e9/volumes" Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.837244 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/cinder-scheduler-0" Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.873482 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-5b5f75f756-jlhwg" event={"ID":"5c82b179-3510-4fb2-9352-d41e7193d221","Type":"ContainerDied","Data":"c278653c52e37dbf15ebd8eb218ca2dea6ad7489d8831b811d37f5ffc16ec9b9"} Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.873536 4967 scope.go:117] "RemoveContainer" containerID="08d7b35bd810dabe97b43264d89fe7cfe49649ec8d7c9897c002dfd2a784d9a5" Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.873660 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-5b5f75f756-jlhwg" Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.880743 4967 generic.go:334] "Generic (PLEG): container finished" podID="eebc2018-61d9-4fe8-ace1-fcb6d817fe80" containerID="b7b9e2f154cda6c793b1118b11625c7241f0fadc381fc945277f06af7eefdc40" exitCode=0 Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.880772 4967 generic.go:334] "Generic (PLEG): container finished" podID="eebc2018-61d9-4fe8-ace1-fcb6d817fe80" containerID="49b331cfb61ba6df00a1aebb22e762b74800f8075ba3aef3dc65388ffee26137" exitCode=2 Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.880782 4967 generic.go:334] "Generic (PLEG): container finished" podID="eebc2018-61d9-4fe8-ace1-fcb6d817fe80" containerID="29a1b90f5cd75564b573dc3d82834afff7c4d05f0f9665c1a7a693e17da9c687" exitCode=0 Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.880835 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eebc2018-61d9-4fe8-ace1-fcb6d817fe80","Type":"ContainerDied","Data":"b7b9e2f154cda6c793b1118b11625c7241f0fadc381fc945277f06af7eefdc40"} Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.880858 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eebc2018-61d9-4fe8-ace1-fcb6d817fe80","Type":"ContainerDied","Data":"49b331cfb61ba6df00a1aebb22e762b74800f8075ba3aef3dc65388ffee26137"} Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.880869 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eebc2018-61d9-4fe8-ace1-fcb6d817fe80","Type":"ContainerDied","Data":"29a1b90f5cd75564b573dc3d82834afff7c4d05f0f9665c1a7a693e17da9c687"} Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.883163 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"662b8e47-db97-4003-892a-48c25cba4589","Type":"ContainerStarted","Data":"873ed5ec64809c56f14048935649e48cc79d4aa6f8ce30c9310e0076831baf73"} Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.899133 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"53708cdd-1297-4302-9d8a-e00ba57c01b3","Type":"ContainerStarted","Data":"f413bc4884166742c957fe92ca25a6009e2f4ce0c57554afb76b7837b916ff97"} Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.905279 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-5b5f75f756-jlhwg"] Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.912739 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ironic-5b5f75f756-jlhwg"] Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.931583 4967 scope.go:117] "RemoveContainer" containerID="565f03d42b75e92dcace4c1ad5111a96bc6b2d2fbcbc46bc88b7137676ce22f6" Oct 11 04:09:00 crc kubenswrapper[4967]: I1011 04:09:00.963354 4967 scope.go:117] "RemoveContainer" containerID="fab633c4576fe35b347aa6b382b58410ac65fa121b618fac8f93e6338a61127b" Oct 11 04:09:01 crc kubenswrapper[4967]: I1011 04:09:01.644767 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:09:01 crc kubenswrapper[4967]: I1011 04:09:01.674144 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-combined-ca-bundle\") pod \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\" (UID: \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\") " Oct 11 04:09:01 crc kubenswrapper[4967]: I1011 04:09:01.683371 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-config-data\") pod \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\" (UID: \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\") " Oct 11 04:09:01 crc kubenswrapper[4967]: I1011 04:09:01.683438 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dh476\" (UniqueName: \"kubernetes.io/projected/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-kube-api-access-dh476\") pod \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\" (UID: \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\") " Oct 11 04:09:01 crc kubenswrapper[4967]: I1011 04:09:01.683506 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-run-httpd\") pod \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\" (UID: \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\") " Oct 11 04:09:01 crc kubenswrapper[4967]: I1011 04:09:01.683538 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-scripts\") pod \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\" (UID: \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\") " Oct 11 04:09:01 crc kubenswrapper[4967]: I1011 04:09:01.683575 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-sg-core-conf-yaml\") pod \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\" (UID: \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\") " Oct 11 04:09:01 crc kubenswrapper[4967]: I1011 04:09:01.684672 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-log-httpd\") pod \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\" (UID: \"eebc2018-61d9-4fe8-ace1-fcb6d817fe80\") " Oct 11 04:09:01 crc kubenswrapper[4967]: I1011 04:09:01.684519 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "eebc2018-61d9-4fe8-ace1-fcb6d817fe80" (UID: "eebc2018-61d9-4fe8-ace1-fcb6d817fe80"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:09:01 crc kubenswrapper[4967]: I1011 04:09:01.686641 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "eebc2018-61d9-4fe8-ace1-fcb6d817fe80" (UID: "eebc2018-61d9-4fe8-ace1-fcb6d817fe80"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:09:01 crc kubenswrapper[4967]: I1011 04:09:01.695350 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-kube-api-access-dh476" (OuterVolumeSpecName: "kube-api-access-dh476") pod "eebc2018-61d9-4fe8-ace1-fcb6d817fe80" (UID: "eebc2018-61d9-4fe8-ace1-fcb6d817fe80"). InnerVolumeSpecName "kube-api-access-dh476". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:09:01 crc kubenswrapper[4967]: I1011 04:09:01.701177 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-scripts" (OuterVolumeSpecName: "scripts") pod "eebc2018-61d9-4fe8-ace1-fcb6d817fe80" (UID: "eebc2018-61d9-4fe8-ace1-fcb6d817fe80"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:09:01 crc kubenswrapper[4967]: I1011 04:09:01.747282 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "eebc2018-61d9-4fe8-ace1-fcb6d817fe80" (UID: "eebc2018-61d9-4fe8-ace1-fcb6d817fe80"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:09:01 crc kubenswrapper[4967]: I1011 04:09:01.789105 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dh476\" (UniqueName: \"kubernetes.io/projected/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-kube-api-access-dh476\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:01 crc kubenswrapper[4967]: I1011 04:09:01.789138 4967 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:01 crc kubenswrapper[4967]: I1011 04:09:01.789151 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:01 crc kubenswrapper[4967]: I1011 04:09:01.789165 4967 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:01 crc kubenswrapper[4967]: I1011 04:09:01.789176 4967 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:01 crc kubenswrapper[4967]: I1011 04:09:01.805084 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "eebc2018-61d9-4fe8-ace1-fcb6d817fe80" (UID: "eebc2018-61d9-4fe8-ace1-fcb6d817fe80"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:09:01 crc kubenswrapper[4967]: I1011 04:09:01.852951 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-config-data" (OuterVolumeSpecName: "config-data") pod "eebc2018-61d9-4fe8-ace1-fcb6d817fe80" (UID: "eebc2018-61d9-4fe8-ace1-fcb6d817fe80"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:09:01 crc kubenswrapper[4967]: I1011 04:09:01.891439 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:01 crc kubenswrapper[4967]: I1011 04:09:01.891467 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/eebc2018-61d9-4fe8-ace1-fcb6d817fe80-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:01 crc kubenswrapper[4967]: I1011 04:09:01.922432 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"662b8e47-db97-4003-892a-48c25cba4589","Type":"ContainerStarted","Data":"ed04a1bb08d87cd4b058371524bc33ef3cc73fbadd714b70ad52656196bad4af"} Oct 11 04:09:01 crc kubenswrapper[4967]: I1011 04:09:01.922493 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-external-api-0" event={"ID":"662b8e47-db97-4003-892a-48c25cba4589","Type":"ContainerStarted","Data":"2447299f316630961fb7b2b3067341f41af9f78255cf6311e1dfc5821f49315e"} Oct 11 04:09:01 crc kubenswrapper[4967]: I1011 04:09:01.926226 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"53708cdd-1297-4302-9d8a-e00ba57c01b3","Type":"ContainerStarted","Data":"19564b715e30659a7c0f5724e28636a53612eac45edc45a9f28b0ee961990993"} Oct 11 04:09:01 crc kubenswrapper[4967]: I1011 04:09:01.931732 4967 generic.go:334] "Generic (PLEG): container finished" podID="eebc2018-61d9-4fe8-ace1-fcb6d817fe80" containerID="c2013cb63bc9eb909af03fc2c4e288a5a06a28fd0bff917277631dba4221e1e2" exitCode=0 Oct 11 04:09:01 crc kubenswrapper[4967]: I1011 04:09:01.931791 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eebc2018-61d9-4fe8-ace1-fcb6d817fe80","Type":"ContainerDied","Data":"c2013cb63bc9eb909af03fc2c4e288a5a06a28fd0bff917277631dba4221e1e2"} Oct 11 04:09:01 crc kubenswrapper[4967]: I1011 04:09:01.931822 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:09:01 crc kubenswrapper[4967]: I1011 04:09:01.931860 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"eebc2018-61d9-4fe8-ace1-fcb6d817fe80","Type":"ContainerDied","Data":"20317c8eb3833736a0a77dafb89a49b6926509f49b3a108af90686e6a209a5e8"} Oct 11 04:09:01 crc kubenswrapper[4967]: I1011 04:09:01.931885 4967 scope.go:117] "RemoveContainer" containerID="b7b9e2f154cda6c793b1118b11625c7241f0fadc381fc945277f06af7eefdc40" Oct 11 04:09:01 crc kubenswrapper[4967]: I1011 04:09:01.951340 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-external-api-0" podStartSLOduration=3.951318357 podStartE2EDuration="3.951318357s" podCreationTimestamp="2025-10-11 04:08:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:09:01.944511336 +0000 UTC m=+1069.907720269" watchObservedRunningTime="2025-10-11 04:09:01.951318357 +0000 UTC m=+1069.914527280" Oct 11 04:09:01 crc kubenswrapper[4967]: I1011 04:09:01.978747 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.000227 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.001228 4967 scope.go:117] "RemoveContainer" containerID="49b331cfb61ba6df00a1aebb22e762b74800f8075ba3aef3dc65388ffee26137" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.003635 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:09:02 crc kubenswrapper[4967]: E1011 04:09:02.005979 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c82b179-3510-4fb2-9352-d41e7193d221" containerName="init" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.006007 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c82b179-3510-4fb2-9352-d41e7193d221" containerName="init" Oct 11 04:09:02 crc kubenswrapper[4967]: E1011 04:09:02.006020 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eebc2018-61d9-4fe8-ace1-fcb6d817fe80" containerName="proxy-httpd" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.006026 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="eebc2018-61d9-4fe8-ace1-fcb6d817fe80" containerName="proxy-httpd" Oct 11 04:09:02 crc kubenswrapper[4967]: E1011 04:09:02.006051 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eebc2018-61d9-4fe8-ace1-fcb6d817fe80" containerName="sg-core" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.006056 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="eebc2018-61d9-4fe8-ace1-fcb6d817fe80" containerName="sg-core" Oct 11 04:09:02 crc kubenswrapper[4967]: E1011 04:09:02.006086 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c82b179-3510-4fb2-9352-d41e7193d221" containerName="ironic-api" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.006092 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c82b179-3510-4fb2-9352-d41e7193d221" containerName="ironic-api" Oct 11 04:09:02 crc kubenswrapper[4967]: E1011 04:09:02.006102 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c82b179-3510-4fb2-9352-d41e7193d221" containerName="ironic-api" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.006108 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c82b179-3510-4fb2-9352-d41e7193d221" containerName="ironic-api" Oct 11 04:09:02 crc kubenswrapper[4967]: E1011 04:09:02.006122 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eebc2018-61d9-4fe8-ace1-fcb6d817fe80" containerName="ceilometer-notification-agent" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.006128 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="eebc2018-61d9-4fe8-ace1-fcb6d817fe80" containerName="ceilometer-notification-agent" Oct 11 04:09:02 crc kubenswrapper[4967]: E1011 04:09:02.006144 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5c82b179-3510-4fb2-9352-d41e7193d221" containerName="ironic-api-log" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.006151 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="5c82b179-3510-4fb2-9352-d41e7193d221" containerName="ironic-api-log" Oct 11 04:09:02 crc kubenswrapper[4967]: E1011 04:09:02.006166 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eebc2018-61d9-4fe8-ace1-fcb6d817fe80" containerName="ceilometer-central-agent" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.006172 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="eebc2018-61d9-4fe8-ace1-fcb6d817fe80" containerName="ceilometer-central-agent" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.006371 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="eebc2018-61d9-4fe8-ace1-fcb6d817fe80" containerName="proxy-httpd" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.006392 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="eebc2018-61d9-4fe8-ace1-fcb6d817fe80" containerName="ceilometer-notification-agent" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.006405 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="eebc2018-61d9-4fe8-ace1-fcb6d817fe80" containerName="ceilometer-central-agent" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.006414 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c82b179-3510-4fb2-9352-d41e7193d221" containerName="ironic-api" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.006422 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c82b179-3510-4fb2-9352-d41e7193d221" containerName="ironic-api-log" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.006433 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="eebc2018-61d9-4fe8-ace1-fcb6d817fe80" containerName="sg-core" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.006442 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="5c82b179-3510-4fb2-9352-d41e7193d221" containerName="ironic-api" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.008058 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.014593 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.018354 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.063578 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.070624 4967 scope.go:117] "RemoveContainer" containerID="29a1b90f5cd75564b573dc3d82834afff7c4d05f0f9665c1a7a693e17da9c687" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.095233 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\") " pod="openstack/ceilometer-0" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.095435 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\") " pod="openstack/ceilometer-0" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.095470 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-run-httpd\") pod \"ceilometer-0\" (UID: \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\") " pod="openstack/ceilometer-0" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.095586 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-log-httpd\") pod \"ceilometer-0\" (UID: \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\") " pod="openstack/ceilometer-0" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.095619 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-scripts\") pod \"ceilometer-0\" (UID: \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\") " pod="openstack/ceilometer-0" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.095654 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mnjrd\" (UniqueName: \"kubernetes.io/projected/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-kube-api-access-mnjrd\") pod \"ceilometer-0\" (UID: \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\") " pod="openstack/ceilometer-0" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.095678 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-config-data\") pod \"ceilometer-0\" (UID: \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\") " pod="openstack/ceilometer-0" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.096049 4967 scope.go:117] "RemoveContainer" containerID="c2013cb63bc9eb909af03fc2c4e288a5a06a28fd0bff917277631dba4221e1e2" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.197154 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-log-httpd\") pod \"ceilometer-0\" (UID: \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\") " pod="openstack/ceilometer-0" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.197207 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-scripts\") pod \"ceilometer-0\" (UID: \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\") " pod="openstack/ceilometer-0" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.197284 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mnjrd\" (UniqueName: \"kubernetes.io/projected/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-kube-api-access-mnjrd\") pod \"ceilometer-0\" (UID: \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\") " pod="openstack/ceilometer-0" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.197313 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-config-data\") pod \"ceilometer-0\" (UID: \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\") " pod="openstack/ceilometer-0" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.197353 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\") " pod="openstack/ceilometer-0" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.197392 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\") " pod="openstack/ceilometer-0" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.197418 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-run-httpd\") pod \"ceilometer-0\" (UID: \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\") " pod="openstack/ceilometer-0" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.197855 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-run-httpd\") pod \"ceilometer-0\" (UID: \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\") " pod="openstack/ceilometer-0" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.198042 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-log-httpd\") pod \"ceilometer-0\" (UID: \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\") " pod="openstack/ceilometer-0" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.203417 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-scripts\") pod \"ceilometer-0\" (UID: \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\") " pod="openstack/ceilometer-0" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.203542 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\") " pod="openstack/ceilometer-0" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.206738 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-config-data\") pod \"ceilometer-0\" (UID: \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\") " pod="openstack/ceilometer-0" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.207253 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\") " pod="openstack/ceilometer-0" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.211991 4967 scope.go:117] "RemoveContainer" containerID="b7b9e2f154cda6c793b1118b11625c7241f0fadc381fc945277f06af7eefdc40" Oct 11 04:09:02 crc kubenswrapper[4967]: E1011 04:09:02.221492 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b7b9e2f154cda6c793b1118b11625c7241f0fadc381fc945277f06af7eefdc40\": container with ID starting with b7b9e2f154cda6c793b1118b11625c7241f0fadc381fc945277f06af7eefdc40 not found: ID does not exist" containerID="b7b9e2f154cda6c793b1118b11625c7241f0fadc381fc945277f06af7eefdc40" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.221574 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b7b9e2f154cda6c793b1118b11625c7241f0fadc381fc945277f06af7eefdc40"} err="failed to get container status \"b7b9e2f154cda6c793b1118b11625c7241f0fadc381fc945277f06af7eefdc40\": rpc error: code = NotFound desc = could not find container \"b7b9e2f154cda6c793b1118b11625c7241f0fadc381fc945277f06af7eefdc40\": container with ID starting with b7b9e2f154cda6c793b1118b11625c7241f0fadc381fc945277f06af7eefdc40 not found: ID does not exist" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.221673 4967 scope.go:117] "RemoveContainer" containerID="49b331cfb61ba6df00a1aebb22e762b74800f8075ba3aef3dc65388ffee26137" Oct 11 04:09:02 crc kubenswrapper[4967]: E1011 04:09:02.222262 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"49b331cfb61ba6df00a1aebb22e762b74800f8075ba3aef3dc65388ffee26137\": container with ID starting with 49b331cfb61ba6df00a1aebb22e762b74800f8075ba3aef3dc65388ffee26137 not found: ID does not exist" containerID="49b331cfb61ba6df00a1aebb22e762b74800f8075ba3aef3dc65388ffee26137" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.222314 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"49b331cfb61ba6df00a1aebb22e762b74800f8075ba3aef3dc65388ffee26137"} err="failed to get container status \"49b331cfb61ba6df00a1aebb22e762b74800f8075ba3aef3dc65388ffee26137\": rpc error: code = NotFound desc = could not find container \"49b331cfb61ba6df00a1aebb22e762b74800f8075ba3aef3dc65388ffee26137\": container with ID starting with 49b331cfb61ba6df00a1aebb22e762b74800f8075ba3aef3dc65388ffee26137 not found: ID does not exist" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.222341 4967 scope.go:117] "RemoveContainer" containerID="29a1b90f5cd75564b573dc3d82834afff7c4d05f0f9665c1a7a693e17da9c687" Oct 11 04:09:02 crc kubenswrapper[4967]: E1011 04:09:02.222595 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"29a1b90f5cd75564b573dc3d82834afff7c4d05f0f9665c1a7a693e17da9c687\": container with ID starting with 29a1b90f5cd75564b573dc3d82834afff7c4d05f0f9665c1a7a693e17da9c687 not found: ID does not exist" containerID="29a1b90f5cd75564b573dc3d82834afff7c4d05f0f9665c1a7a693e17da9c687" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.222624 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"29a1b90f5cd75564b573dc3d82834afff7c4d05f0f9665c1a7a693e17da9c687"} err="failed to get container status \"29a1b90f5cd75564b573dc3d82834afff7c4d05f0f9665c1a7a693e17da9c687\": rpc error: code = NotFound desc = could not find container \"29a1b90f5cd75564b573dc3d82834afff7c4d05f0f9665c1a7a693e17da9c687\": container with ID starting with 29a1b90f5cd75564b573dc3d82834afff7c4d05f0f9665c1a7a693e17da9c687 not found: ID does not exist" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.222637 4967 scope.go:117] "RemoveContainer" containerID="c2013cb63bc9eb909af03fc2c4e288a5a06a28fd0bff917277631dba4221e1e2" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.222723 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mnjrd\" (UniqueName: \"kubernetes.io/projected/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-kube-api-access-mnjrd\") pod \"ceilometer-0\" (UID: \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\") " pod="openstack/ceilometer-0" Oct 11 04:09:02 crc kubenswrapper[4967]: E1011 04:09:02.223004 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2013cb63bc9eb909af03fc2c4e288a5a06a28fd0bff917277631dba4221e1e2\": container with ID starting with c2013cb63bc9eb909af03fc2c4e288a5a06a28fd0bff917277631dba4221e1e2 not found: ID does not exist" containerID="c2013cb63bc9eb909af03fc2c4e288a5a06a28fd0bff917277631dba4221e1e2" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.223030 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2013cb63bc9eb909af03fc2c4e288a5a06a28fd0bff917277631dba4221e1e2"} err="failed to get container status \"c2013cb63bc9eb909af03fc2c4e288a5a06a28fd0bff917277631dba4221e1e2\": rpc error: code = NotFound desc = could not find container \"c2013cb63bc9eb909af03fc2c4e288a5a06a28fd0bff917277631dba4221e1e2\": container with ID starting with c2013cb63bc9eb909af03fc2c4e288a5a06a28fd0bff917277631dba4221e1e2 not found: ID does not exist" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.363551 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.838792 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5c82b179-3510-4fb2-9352-d41e7193d221" path="/var/lib/kubelet/pods/5c82b179-3510-4fb2-9352-d41e7193d221/volumes" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.840217 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="eebc2018-61d9-4fe8-ace1-fcb6d817fe80" path="/var/lib/kubelet/pods/eebc2018-61d9-4fe8-ace1-fcb6d817fe80/volumes" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.894118 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-652d-account-create-djxcs"] Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.895640 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-652d-account-create-djxcs" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.898472 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-db-secret" Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.914840 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.943082 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-652d-account-create-djxcs"] Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.956211 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/glance-default-internal-api-0" event={"ID":"53708cdd-1297-4302-9d8a-e00ba57c01b3","Type":"ContainerStarted","Data":"1e63b7a73e6ae329df14f9b6c4d8d0b7d8f005fd584d3a6ff30bc2aa5e8eae65"} Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.965253 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33","Type":"ContainerStarted","Data":"b9354f9bcc0abc860f68c16c5003b0e0deec8a4d734611fc7bcf833daa875149"} Oct 11 04:09:02 crc kubenswrapper[4967]: I1011 04:09:02.984976 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/glance-default-internal-api-0" podStartSLOduration=3.984953601 podStartE2EDuration="3.984953601s" podCreationTimestamp="2025-10-11 04:08:59 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:09:02.975180848 +0000 UTC m=+1070.938389781" watchObservedRunningTime="2025-10-11 04:09:02.984953601 +0000 UTC m=+1070.948162534" Oct 11 04:09:03 crc kubenswrapper[4967]: I1011 04:09:03.019634 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzwgz\" (UniqueName: \"kubernetes.io/projected/917decdc-0498-4bf6-a45f-cba451199483-kube-api-access-fzwgz\") pod \"nova-cell0-652d-account-create-djxcs\" (UID: \"917decdc-0498-4bf6-a45f-cba451199483\") " pod="openstack/nova-cell0-652d-account-create-djxcs" Oct 11 04:09:03 crc kubenswrapper[4967]: I1011 04:09:03.122353 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzwgz\" (UniqueName: \"kubernetes.io/projected/917decdc-0498-4bf6-a45f-cba451199483-kube-api-access-fzwgz\") pod \"nova-cell0-652d-account-create-djxcs\" (UID: \"917decdc-0498-4bf6-a45f-cba451199483\") " pod="openstack/nova-cell0-652d-account-create-djxcs" Oct 11 04:09:03 crc kubenswrapper[4967]: I1011 04:09:03.149443 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzwgz\" (UniqueName: \"kubernetes.io/projected/917decdc-0498-4bf6-a45f-cba451199483-kube-api-access-fzwgz\") pod \"nova-cell0-652d-account-create-djxcs\" (UID: \"917decdc-0498-4bf6-a45f-cba451199483\") " pod="openstack/nova-cell0-652d-account-create-djxcs" Oct 11 04:09:03 crc kubenswrapper[4967]: I1011 04:09:03.223645 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-652d-account-create-djxcs" Oct 11 04:09:03 crc kubenswrapper[4967]: I1011 04:09:03.707572 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-652d-account-create-djxcs"] Oct 11 04:09:03 crc kubenswrapper[4967]: W1011 04:09:03.710150 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod917decdc_0498_4bf6_a45f_cba451199483.slice/crio-628a3b6f2cfbcf2589fd7f5501d196fdac6b61477a334a894b9fde3f6720d06b WatchSource:0}: Error finding container 628a3b6f2cfbcf2589fd7f5501d196fdac6b61477a334a894b9fde3f6720d06b: Status 404 returned error can't find the container with id 628a3b6f2cfbcf2589fd7f5501d196fdac6b61477a334a894b9fde3f6720d06b Oct 11 04:09:03 crc kubenswrapper[4967]: I1011 04:09:03.976837 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33","Type":"ContainerStarted","Data":"61206f1f40a74f25942e30bbb070ab3e34766c2f1fe78d0ead1571eb12628643"} Oct 11 04:09:03 crc kubenswrapper[4967]: I1011 04:09:03.979137 4967 generic.go:334] "Generic (PLEG): container finished" podID="917decdc-0498-4bf6-a45f-cba451199483" containerID="478a3e8f1b62c6faf971349d240c3f760ef34f0e4761a46e75a50421e9b3abe6" exitCode=0 Oct 11 04:09:03 crc kubenswrapper[4967]: I1011 04:09:03.980340 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-652d-account-create-djxcs" event={"ID":"917decdc-0498-4bf6-a45f-cba451199483","Type":"ContainerDied","Data":"478a3e8f1b62c6faf971349d240c3f760ef34f0e4761a46e75a50421e9b3abe6"} Oct 11 04:09:03 crc kubenswrapper[4967]: I1011 04:09:03.980417 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-652d-account-create-djxcs" event={"ID":"917decdc-0498-4bf6-a45f-cba451199483","Type":"ContainerStarted","Data":"628a3b6f2cfbcf2589fd7f5501d196fdac6b61477a334a894b9fde3f6720d06b"} Oct 11 04:09:04 crc kubenswrapper[4967]: I1011 04:09:04.989925 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33","Type":"ContainerStarted","Data":"5e00aad7c84c286e780dce5ac301bab981272537243892e81da642ce38e7a6ce"} Oct 11 04:09:04 crc kubenswrapper[4967]: I1011 04:09:04.990695 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33","Type":"ContainerStarted","Data":"ba24fdb99bd4eb58c090503135a2ffba357961da62cd7632202ce6f95a43418e"} Oct 11 04:09:05 crc kubenswrapper[4967]: I1011 04:09:05.374446 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-652d-account-create-djxcs" Oct 11 04:09:05 crc kubenswrapper[4967]: I1011 04:09:05.472181 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzwgz\" (UniqueName: \"kubernetes.io/projected/917decdc-0498-4bf6-a45f-cba451199483-kube-api-access-fzwgz\") pod \"917decdc-0498-4bf6-a45f-cba451199483\" (UID: \"917decdc-0498-4bf6-a45f-cba451199483\") " Oct 11 04:09:05 crc kubenswrapper[4967]: I1011 04:09:05.490811 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/917decdc-0498-4bf6-a45f-cba451199483-kube-api-access-fzwgz" (OuterVolumeSpecName: "kube-api-access-fzwgz") pod "917decdc-0498-4bf6-a45f-cba451199483" (UID: "917decdc-0498-4bf6-a45f-cba451199483"). InnerVolumeSpecName "kube-api-access-fzwgz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:09:05 crc kubenswrapper[4967]: I1011 04:09:05.574887 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzwgz\" (UniqueName: \"kubernetes.io/projected/917decdc-0498-4bf6-a45f-cba451199483-kube-api-access-fzwgz\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:06 crc kubenswrapper[4967]: I1011 04:09:06.000728 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-652d-account-create-djxcs" event={"ID":"917decdc-0498-4bf6-a45f-cba451199483","Type":"ContainerDied","Data":"628a3b6f2cfbcf2589fd7f5501d196fdac6b61477a334a894b9fde3f6720d06b"} Oct 11 04:09:06 crc kubenswrapper[4967]: I1011 04:09:06.000785 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="628a3b6f2cfbcf2589fd7f5501d196fdac6b61477a334a894b9fde3f6720d06b" Oct 11 04:09:06 crc kubenswrapper[4967]: I1011 04:09:06.000855 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-652d-account-create-djxcs" Oct 11 04:09:06 crc kubenswrapper[4967]: I1011 04:09:06.053912 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/cinder-scheduler-0" Oct 11 04:09:07 crc kubenswrapper[4967]: I1011 04:09:07.012700 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33","Type":"ContainerStarted","Data":"894b5ed10218944033f0d21b5f1591ad9b5c3e51a9d40368be249f1e22cb1c07"} Oct 11 04:09:07 crc kubenswrapper[4967]: I1011 04:09:07.013108 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 11 04:09:07 crc kubenswrapper[4967]: I1011 04:09:07.043153 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.6589750260000002 podStartE2EDuration="6.0431328s" podCreationTimestamp="2025-10-11 04:09:01 +0000 UTC" firstStartedPulling="2025-10-11 04:09:02.924881965 +0000 UTC m=+1070.888090898" lastFinishedPulling="2025-10-11 04:09:06.309039739 +0000 UTC m=+1074.272248672" observedRunningTime="2025-10-11 04:09:07.031810178 +0000 UTC m=+1074.995019151" watchObservedRunningTime="2025-10-11 04:09:07.0431328 +0000 UTC m=+1075.006341753" Oct 11 04:09:07 crc kubenswrapper[4967]: I1011 04:09:07.984311 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xh6jw"] Oct 11 04:09:07 crc kubenswrapper[4967]: E1011 04:09:07.984677 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="917decdc-0498-4bf6-a45f-cba451199483" containerName="mariadb-account-create" Oct 11 04:09:07 crc kubenswrapper[4967]: I1011 04:09:07.984695 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="917decdc-0498-4bf6-a45f-cba451199483" containerName="mariadb-account-create" Oct 11 04:09:07 crc kubenswrapper[4967]: I1011 04:09:07.984884 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="917decdc-0498-4bf6-a45f-cba451199483" containerName="mariadb-account-create" Oct 11 04:09:07 crc kubenswrapper[4967]: I1011 04:09:07.985465 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xh6jw" Oct 11 04:09:07 crc kubenswrapper[4967]: I1011 04:09:07.989668 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-scripts" Oct 11 04:09:07 crc kubenswrapper[4967]: I1011 04:09:07.989832 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-fj2pf" Oct 11 04:09:07 crc kubenswrapper[4967]: I1011 04:09:07.990264 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 11 04:09:08 crc kubenswrapper[4967]: I1011 04:09:08.001590 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xh6jw"] Oct 11 04:09:08 crc kubenswrapper[4967]: I1011 04:09:08.024772 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4a234ae-b210-4ac3-9bf2-841f586ee4a3-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xh6jw\" (UID: \"d4a234ae-b210-4ac3-9bf2-841f586ee4a3\") " pod="openstack/nova-cell0-conductor-db-sync-xh6jw" Oct 11 04:09:08 crc kubenswrapper[4967]: I1011 04:09:08.024927 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4a234ae-b210-4ac3-9bf2-841f586ee4a3-scripts\") pod \"nova-cell0-conductor-db-sync-xh6jw\" (UID: \"d4a234ae-b210-4ac3-9bf2-841f586ee4a3\") " pod="openstack/nova-cell0-conductor-db-sync-xh6jw" Oct 11 04:09:08 crc kubenswrapper[4967]: I1011 04:09:08.024981 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4a234ae-b210-4ac3-9bf2-841f586ee4a3-config-data\") pod \"nova-cell0-conductor-db-sync-xh6jw\" (UID: \"d4a234ae-b210-4ac3-9bf2-841f586ee4a3\") " pod="openstack/nova-cell0-conductor-db-sync-xh6jw" Oct 11 04:09:08 crc kubenswrapper[4967]: I1011 04:09:08.025156 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bxp55\" (UniqueName: \"kubernetes.io/projected/d4a234ae-b210-4ac3-9bf2-841f586ee4a3-kube-api-access-bxp55\") pod \"nova-cell0-conductor-db-sync-xh6jw\" (UID: \"d4a234ae-b210-4ac3-9bf2-841f586ee4a3\") " pod="openstack/nova-cell0-conductor-db-sync-xh6jw" Oct 11 04:09:08 crc kubenswrapper[4967]: I1011 04:09:08.127024 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4a234ae-b210-4ac3-9bf2-841f586ee4a3-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xh6jw\" (UID: \"d4a234ae-b210-4ac3-9bf2-841f586ee4a3\") " pod="openstack/nova-cell0-conductor-db-sync-xh6jw" Oct 11 04:09:08 crc kubenswrapper[4967]: I1011 04:09:08.127190 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4a234ae-b210-4ac3-9bf2-841f586ee4a3-scripts\") pod \"nova-cell0-conductor-db-sync-xh6jw\" (UID: \"d4a234ae-b210-4ac3-9bf2-841f586ee4a3\") " pod="openstack/nova-cell0-conductor-db-sync-xh6jw" Oct 11 04:09:08 crc kubenswrapper[4967]: I1011 04:09:08.127235 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4a234ae-b210-4ac3-9bf2-841f586ee4a3-config-data\") pod \"nova-cell0-conductor-db-sync-xh6jw\" (UID: \"d4a234ae-b210-4ac3-9bf2-841f586ee4a3\") " pod="openstack/nova-cell0-conductor-db-sync-xh6jw" Oct 11 04:09:08 crc kubenswrapper[4967]: I1011 04:09:08.127304 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bxp55\" (UniqueName: \"kubernetes.io/projected/d4a234ae-b210-4ac3-9bf2-841f586ee4a3-kube-api-access-bxp55\") pod \"nova-cell0-conductor-db-sync-xh6jw\" (UID: \"d4a234ae-b210-4ac3-9bf2-841f586ee4a3\") " pod="openstack/nova-cell0-conductor-db-sync-xh6jw" Oct 11 04:09:08 crc kubenswrapper[4967]: I1011 04:09:08.135646 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4a234ae-b210-4ac3-9bf2-841f586ee4a3-scripts\") pod \"nova-cell0-conductor-db-sync-xh6jw\" (UID: \"d4a234ae-b210-4ac3-9bf2-841f586ee4a3\") " pod="openstack/nova-cell0-conductor-db-sync-xh6jw" Oct 11 04:09:08 crc kubenswrapper[4967]: I1011 04:09:08.135835 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4a234ae-b210-4ac3-9bf2-841f586ee4a3-combined-ca-bundle\") pod \"nova-cell0-conductor-db-sync-xh6jw\" (UID: \"d4a234ae-b210-4ac3-9bf2-841f586ee4a3\") " pod="openstack/nova-cell0-conductor-db-sync-xh6jw" Oct 11 04:09:08 crc kubenswrapper[4967]: I1011 04:09:08.135849 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4a234ae-b210-4ac3-9bf2-841f586ee4a3-config-data\") pod \"nova-cell0-conductor-db-sync-xh6jw\" (UID: \"d4a234ae-b210-4ac3-9bf2-841f586ee4a3\") " pod="openstack/nova-cell0-conductor-db-sync-xh6jw" Oct 11 04:09:08 crc kubenswrapper[4967]: I1011 04:09:08.146620 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bxp55\" (UniqueName: \"kubernetes.io/projected/d4a234ae-b210-4ac3-9bf2-841f586ee4a3-kube-api-access-bxp55\") pod \"nova-cell0-conductor-db-sync-xh6jw\" (UID: \"d4a234ae-b210-4ac3-9bf2-841f586ee4a3\") " pod="openstack/nova-cell0-conductor-db-sync-xh6jw" Oct 11 04:09:08 crc kubenswrapper[4967]: I1011 04:09:08.325577 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xh6jw" Oct 11 04:09:08 crc kubenswrapper[4967]: I1011 04:09:08.865268 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xh6jw"] Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.041036 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xh6jw" event={"ID":"d4a234ae-b210-4ac3-9bf2-841f586ee4a3","Type":"ContainerStarted","Data":"5a2c3abcc638399339b562559ac6368d2e3d72c071d2391fc2242f4807c68e60"} Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.071828 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-inspector-db-sync-m5k2b"] Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.072981 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-sync-m5k2b" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.075488 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-scripts" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.075612 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-config-data" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.087985 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-db-sync-m5k2b"] Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.147999 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qj465\" (UniqueName: \"kubernetes.io/projected/1d9749c8-f6a4-497f-b6dd-794865243379-kube-api-access-qj465\") pod \"ironic-inspector-db-sync-m5k2b\" (UID: \"1d9749c8-f6a4-497f-b6dd-794865243379\") " pod="openstack/ironic-inspector-db-sync-m5k2b" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.148054 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/1d9749c8-f6a4-497f-b6dd-794865243379-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-db-sync-m5k2b\" (UID: \"1d9749c8-f6a4-497f-b6dd-794865243379\") " pod="openstack/ironic-inspector-db-sync-m5k2b" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.148104 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d9749c8-f6a4-497f-b6dd-794865243379-scripts\") pod \"ironic-inspector-db-sync-m5k2b\" (UID: \"1d9749c8-f6a4-497f-b6dd-794865243379\") " pod="openstack/ironic-inspector-db-sync-m5k2b" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.148185 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1d9749c8-f6a4-497f-b6dd-794865243379-config\") pod \"ironic-inspector-db-sync-m5k2b\" (UID: \"1d9749c8-f6a4-497f-b6dd-794865243379\") " pod="openstack/ironic-inspector-db-sync-m5k2b" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.148285 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/1d9749c8-f6a4-497f-b6dd-794865243379-var-lib-ironic\") pod \"ironic-inspector-db-sync-m5k2b\" (UID: \"1d9749c8-f6a4-497f-b6dd-794865243379\") " pod="openstack/ironic-inspector-db-sync-m5k2b" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.148305 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/1d9749c8-f6a4-497f-b6dd-794865243379-etc-podinfo\") pod \"ironic-inspector-db-sync-m5k2b\" (UID: \"1d9749c8-f6a4-497f-b6dd-794865243379\") " pod="openstack/ironic-inspector-db-sync-m5k2b" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.148329 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d9749c8-f6a4-497f-b6dd-794865243379-combined-ca-bundle\") pod \"ironic-inspector-db-sync-m5k2b\" (UID: \"1d9749c8-f6a4-497f-b6dd-794865243379\") " pod="openstack/ironic-inspector-db-sync-m5k2b" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.249989 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/1d9749c8-f6a4-497f-b6dd-794865243379-etc-podinfo\") pod \"ironic-inspector-db-sync-m5k2b\" (UID: \"1d9749c8-f6a4-497f-b6dd-794865243379\") " pod="openstack/ironic-inspector-db-sync-m5k2b" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.250442 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/1d9749c8-f6a4-497f-b6dd-794865243379-var-lib-ironic\") pod \"ironic-inspector-db-sync-m5k2b\" (UID: \"1d9749c8-f6a4-497f-b6dd-794865243379\") " pod="openstack/ironic-inspector-db-sync-m5k2b" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.250483 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d9749c8-f6a4-497f-b6dd-794865243379-combined-ca-bundle\") pod \"ironic-inspector-db-sync-m5k2b\" (UID: \"1d9749c8-f6a4-497f-b6dd-794865243379\") " pod="openstack/ironic-inspector-db-sync-m5k2b" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.250556 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qj465\" (UniqueName: \"kubernetes.io/projected/1d9749c8-f6a4-497f-b6dd-794865243379-kube-api-access-qj465\") pod \"ironic-inspector-db-sync-m5k2b\" (UID: \"1d9749c8-f6a4-497f-b6dd-794865243379\") " pod="openstack/ironic-inspector-db-sync-m5k2b" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.250594 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/1d9749c8-f6a4-497f-b6dd-794865243379-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-db-sync-m5k2b\" (UID: \"1d9749c8-f6a4-497f-b6dd-794865243379\") " pod="openstack/ironic-inspector-db-sync-m5k2b" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.250625 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d9749c8-f6a4-497f-b6dd-794865243379-scripts\") pod \"ironic-inspector-db-sync-m5k2b\" (UID: \"1d9749c8-f6a4-497f-b6dd-794865243379\") " pod="openstack/ironic-inspector-db-sync-m5k2b" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.250665 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1d9749c8-f6a4-497f-b6dd-794865243379-config\") pod \"ironic-inspector-db-sync-m5k2b\" (UID: \"1d9749c8-f6a4-497f-b6dd-794865243379\") " pod="openstack/ironic-inspector-db-sync-m5k2b" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.251047 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/1d9749c8-f6a4-497f-b6dd-794865243379-var-lib-ironic\") pod \"ironic-inspector-db-sync-m5k2b\" (UID: \"1d9749c8-f6a4-497f-b6dd-794865243379\") " pod="openstack/ironic-inspector-db-sync-m5k2b" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.251151 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/1d9749c8-f6a4-497f-b6dd-794865243379-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-db-sync-m5k2b\" (UID: \"1d9749c8-f6a4-497f-b6dd-794865243379\") " pod="openstack/ironic-inspector-db-sync-m5k2b" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.258417 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/1d9749c8-f6a4-497f-b6dd-794865243379-etc-podinfo\") pod \"ironic-inspector-db-sync-m5k2b\" (UID: \"1d9749c8-f6a4-497f-b6dd-794865243379\") " pod="openstack/ironic-inspector-db-sync-m5k2b" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.258465 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d9749c8-f6a4-497f-b6dd-794865243379-scripts\") pod \"ironic-inspector-db-sync-m5k2b\" (UID: \"1d9749c8-f6a4-497f-b6dd-794865243379\") " pod="openstack/ironic-inspector-db-sync-m5k2b" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.263879 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/1d9749c8-f6a4-497f-b6dd-794865243379-config\") pod \"ironic-inspector-db-sync-m5k2b\" (UID: \"1d9749c8-f6a4-497f-b6dd-794865243379\") " pod="openstack/ironic-inspector-db-sync-m5k2b" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.266361 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d9749c8-f6a4-497f-b6dd-794865243379-combined-ca-bundle\") pod \"ironic-inspector-db-sync-m5k2b\" (UID: \"1d9749c8-f6a4-497f-b6dd-794865243379\") " pod="openstack/ironic-inspector-db-sync-m5k2b" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.268794 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qj465\" (UniqueName: \"kubernetes.io/projected/1d9749c8-f6a4-497f-b6dd-794865243379-kube-api-access-qj465\") pod \"ironic-inspector-db-sync-m5k2b\" (UID: \"1d9749c8-f6a4-497f-b6dd-794865243379\") " pod="openstack/ironic-inspector-db-sync-m5k2b" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.390377 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-sync-m5k2b" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.403888 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.403933 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-external-api-0" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.439652 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.460760 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-external-api-0" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.671475 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.671812 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/glance-default-internal-api-0" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.730404 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.734841 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/glance-default-internal-api-0" Oct 11 04:09:09 crc kubenswrapper[4967]: I1011 04:09:09.936842 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-db-sync-m5k2b"] Oct 11 04:09:09 crc kubenswrapper[4967]: W1011 04:09:09.943775 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1d9749c8_f6a4_497f_b6dd_794865243379.slice/crio-19e8690caa331f8312237a061989f958f1bd1b07c3a4204df852d5e5014144f3 WatchSource:0}: Error finding container 19e8690caa331f8312237a061989f958f1bd1b07c3a4204df852d5e5014144f3: Status 404 returned error can't find the container with id 19e8690caa331f8312237a061989f958f1bd1b07c3a4204df852d5e5014144f3 Oct 11 04:09:10 crc kubenswrapper[4967]: I1011 04:09:10.063465 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-sync-m5k2b" event={"ID":"1d9749c8-f6a4-497f-b6dd-794865243379","Type":"ContainerStarted","Data":"19e8690caa331f8312237a061989f958f1bd1b07c3a4204df852d5e5014144f3"} Oct 11 04:09:10 crc kubenswrapper[4967]: I1011 04:09:10.067928 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 11 04:09:10 crc kubenswrapper[4967]: I1011 04:09:10.067976 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 11 04:09:10 crc kubenswrapper[4967]: I1011 04:09:10.067991 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-external-api-0" Oct 11 04:09:10 crc kubenswrapper[4967]: I1011 04:09:10.068003 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/glance-default-internal-api-0" Oct 11 04:09:12 crc kubenswrapper[4967]: I1011 04:09:12.082640 4967 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 11 04:09:12 crc kubenswrapper[4967]: I1011 04:09:12.082869 4967 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 11 04:09:12 crc kubenswrapper[4967]: I1011 04:09:12.083602 4967 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 11 04:09:12 crc kubenswrapper[4967]: I1011 04:09:12.083611 4967 prober_manager.go:312] "Failed to trigger a manual run" probe="Readiness" Oct 11 04:09:12 crc kubenswrapper[4967]: I1011 04:09:12.120416 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 11 04:09:12 crc kubenswrapper[4967]: I1011 04:09:12.120500 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-external-api-0" Oct 11 04:09:12 crc kubenswrapper[4967]: I1011 04:09:12.360273 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 11 04:09:12 crc kubenswrapper[4967]: I1011 04:09:12.364651 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/glance-default-internal-api-0" Oct 11 04:09:12 crc kubenswrapper[4967]: I1011 04:09:12.681377 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-abbb-account-create-szgg7"] Oct 11 04:09:12 crc kubenswrapper[4967]: I1011 04:09:12.687808 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-abbb-account-create-szgg7"] Oct 11 04:09:12 crc kubenswrapper[4967]: I1011 04:09:12.687962 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-abbb-account-create-szgg7" Oct 11 04:09:12 crc kubenswrapper[4967]: I1011 04:09:12.690805 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 11 04:09:12 crc kubenswrapper[4967]: I1011 04:09:12.824206 4967 scope.go:117] "RemoveContainer" containerID="b8c1b1f153783c154d8166fe2c821e9aac2b99008b9620c42dc8206811d3e2aa" Oct 11 04:09:12 crc kubenswrapper[4967]: I1011 04:09:12.826306 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjjtr\" (UniqueName: \"kubernetes.io/projected/cb662b34-1e00-4cbb-a989-347e80fe9886-kube-api-access-sjjtr\") pod \"nova-api-abbb-account-create-szgg7\" (UID: \"cb662b34-1e00-4cbb-a989-347e80fe9886\") " pod="openstack/nova-api-abbb-account-create-szgg7" Oct 11 04:09:12 crc kubenswrapper[4967]: I1011 04:09:12.928237 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjjtr\" (UniqueName: \"kubernetes.io/projected/cb662b34-1e00-4cbb-a989-347e80fe9886-kube-api-access-sjjtr\") pod \"nova-api-abbb-account-create-szgg7\" (UID: \"cb662b34-1e00-4cbb-a989-347e80fe9886\") " pod="openstack/nova-api-abbb-account-create-szgg7" Oct 11 04:09:12 crc kubenswrapper[4967]: I1011 04:09:12.953806 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjjtr\" (UniqueName: \"kubernetes.io/projected/cb662b34-1e00-4cbb-a989-347e80fe9886-kube-api-access-sjjtr\") pod \"nova-api-abbb-account-create-szgg7\" (UID: \"cb662b34-1e00-4cbb-a989-347e80fe9886\") " pod="openstack/nova-api-abbb-account-create-szgg7" Oct 11 04:09:12 crc kubenswrapper[4967]: I1011 04:09:12.982196 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-1117-account-create-76qc9"] Oct 11 04:09:12 crc kubenswrapper[4967]: I1011 04:09:12.983708 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1117-account-create-76qc9" Oct 11 04:09:12 crc kubenswrapper[4967]: I1011 04:09:12.986155 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-db-secret" Oct 11 04:09:12 crc kubenswrapper[4967]: I1011 04:09:12.990433 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-1117-account-create-76qc9"] Oct 11 04:09:13 crc kubenswrapper[4967]: I1011 04:09:13.026476 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-abbb-account-create-szgg7" Oct 11 04:09:13 crc kubenswrapper[4967]: I1011 04:09:13.132741 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tcwld\" (UniqueName: \"kubernetes.io/projected/a758d270-0d47-415f-bae1-ae7288e90245-kube-api-access-tcwld\") pod \"nova-cell1-1117-account-create-76qc9\" (UID: \"a758d270-0d47-415f-bae1-ae7288e90245\") " pod="openstack/nova-cell1-1117-account-create-76qc9" Oct 11 04:09:13 crc kubenswrapper[4967]: I1011 04:09:13.234599 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tcwld\" (UniqueName: \"kubernetes.io/projected/a758d270-0d47-415f-bae1-ae7288e90245-kube-api-access-tcwld\") pod \"nova-cell1-1117-account-create-76qc9\" (UID: \"a758d270-0d47-415f-bae1-ae7288e90245\") " pod="openstack/nova-cell1-1117-account-create-76qc9" Oct 11 04:09:13 crc kubenswrapper[4967]: I1011 04:09:13.255705 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tcwld\" (UniqueName: \"kubernetes.io/projected/a758d270-0d47-415f-bae1-ae7288e90245-kube-api-access-tcwld\") pod \"nova-cell1-1117-account-create-76qc9\" (UID: \"a758d270-0d47-415f-bae1-ae7288e90245\") " pod="openstack/nova-cell1-1117-account-create-76qc9" Oct 11 04:09:13 crc kubenswrapper[4967]: I1011 04:09:13.485233 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1117-account-create-76qc9" Oct 11 04:09:13 crc kubenswrapper[4967]: I1011 04:09:13.536845 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-abbb-account-create-szgg7"] Oct 11 04:09:13 crc kubenswrapper[4967]: I1011 04:09:13.566150 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-db-secret" Oct 11 04:09:13 crc kubenswrapper[4967]: I1011 04:09:13.993290 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-1117-account-create-76qc9"] Oct 11 04:09:14 crc kubenswrapper[4967]: I1011 04:09:14.088866 4967 patch_prober.go:28] interesting pod/machine-config-daemon-sgrv2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:09:14 crc kubenswrapper[4967]: I1011 04:09:14.088920 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:09:14 crc kubenswrapper[4967]: I1011 04:09:14.106610 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-1117-account-create-76qc9" event={"ID":"a758d270-0d47-415f-bae1-ae7288e90245","Type":"ContainerStarted","Data":"e02ce7ce113fc0118d9d04446289357d86af11dad8919dc24c32d14c07df80ea"} Oct 11 04:09:14 crc kubenswrapper[4967]: I1011 04:09:14.110521 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" event={"ID":"4390a950-7185-4fc0-b152-b5927940032b","Type":"ContainerStarted","Data":"93bc6f25445cba6d1e4f70f9b3f209e2f24e716693d78b9c1971de07282202ed"} Oct 11 04:09:14 crc kubenswrapper[4967]: I1011 04:09:14.110851 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" Oct 11 04:09:14 crc kubenswrapper[4967]: I1011 04:09:14.115169 4967 generic.go:334] "Generic (PLEG): container finished" podID="cb662b34-1e00-4cbb-a989-347e80fe9886" containerID="dc95140e2858d0bdbe1c8a9b46e69c6ce135d99beee50963bbfde3504a87caf6" exitCode=0 Oct 11 04:09:14 crc kubenswrapper[4967]: I1011 04:09:14.116177 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-abbb-account-create-szgg7" event={"ID":"cb662b34-1e00-4cbb-a989-347e80fe9886","Type":"ContainerDied","Data":"dc95140e2858d0bdbe1c8a9b46e69c6ce135d99beee50963bbfde3504a87caf6"} Oct 11 04:09:14 crc kubenswrapper[4967]: I1011 04:09:14.116211 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-abbb-account-create-szgg7" event={"ID":"cb662b34-1e00-4cbb-a989-347e80fe9886","Type":"ContainerStarted","Data":"dca4847011c32bd948b1f4642bbff14bbfb4ae3c20a91bdba54061dae653f3c8"} Oct 11 04:09:15 crc kubenswrapper[4967]: I1011 04:09:15.126996 4967 generic.go:334] "Generic (PLEG): container finished" podID="a758d270-0d47-415f-bae1-ae7288e90245" containerID="5c3a8c7a53f8a0661cc2d27adc0254e52ee01bdf4383ead7500f920e6d5de817" exitCode=0 Oct 11 04:09:15 crc kubenswrapper[4967]: I1011 04:09:15.127153 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-1117-account-create-76qc9" event={"ID":"a758d270-0d47-415f-bae1-ae7288e90245","Type":"ContainerDied","Data":"5c3a8c7a53f8a0661cc2d27adc0254e52ee01bdf4383ead7500f920e6d5de817"} Oct 11 04:09:16 crc kubenswrapper[4967]: I1011 04:09:16.066449 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-abbb-account-create-szgg7" Oct 11 04:09:16 crc kubenswrapper[4967]: I1011 04:09:16.144602 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-abbb-account-create-szgg7" Oct 11 04:09:16 crc kubenswrapper[4967]: I1011 04:09:16.144593 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-abbb-account-create-szgg7" event={"ID":"cb662b34-1e00-4cbb-a989-347e80fe9886","Type":"ContainerDied","Data":"dca4847011c32bd948b1f4642bbff14bbfb4ae3c20a91bdba54061dae653f3c8"} Oct 11 04:09:16 crc kubenswrapper[4967]: I1011 04:09:16.144661 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dca4847011c32bd948b1f4642bbff14bbfb4ae3c20a91bdba54061dae653f3c8" Oct 11 04:09:16 crc kubenswrapper[4967]: I1011 04:09:16.192927 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sjjtr\" (UniqueName: \"kubernetes.io/projected/cb662b34-1e00-4cbb-a989-347e80fe9886-kube-api-access-sjjtr\") pod \"cb662b34-1e00-4cbb-a989-347e80fe9886\" (UID: \"cb662b34-1e00-4cbb-a989-347e80fe9886\") " Oct 11 04:09:16 crc kubenswrapper[4967]: I1011 04:09:16.199022 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cb662b34-1e00-4cbb-a989-347e80fe9886-kube-api-access-sjjtr" (OuterVolumeSpecName: "kube-api-access-sjjtr") pod "cb662b34-1e00-4cbb-a989-347e80fe9886" (UID: "cb662b34-1e00-4cbb-a989-347e80fe9886"). InnerVolumeSpecName "kube-api-access-sjjtr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:09:16 crc kubenswrapper[4967]: I1011 04:09:16.295258 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sjjtr\" (UniqueName: \"kubernetes.io/projected/cb662b34-1e00-4cbb-a989-347e80fe9886-kube-api-access-sjjtr\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:17 crc kubenswrapper[4967]: I1011 04:09:17.156264 4967 generic.go:334] "Generic (PLEG): container finished" podID="4390a950-7185-4fc0-b152-b5927940032b" containerID="93bc6f25445cba6d1e4f70f9b3f209e2f24e716693d78b9c1971de07282202ed" exitCode=1 Oct 11 04:09:17 crc kubenswrapper[4967]: I1011 04:09:17.156438 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" event={"ID":"4390a950-7185-4fc0-b152-b5927940032b","Type":"ContainerDied","Data":"93bc6f25445cba6d1e4f70f9b3f209e2f24e716693d78b9c1971de07282202ed"} Oct 11 04:09:17 crc kubenswrapper[4967]: I1011 04:09:17.156590 4967 scope.go:117] "RemoveContainer" containerID="b8c1b1f153783c154d8166fe2c821e9aac2b99008b9620c42dc8206811d3e2aa" Oct 11 04:09:17 crc kubenswrapper[4967]: I1011 04:09:17.157266 4967 scope.go:117] "RemoveContainer" containerID="93bc6f25445cba6d1e4f70f9b3f209e2f24e716693d78b9c1971de07282202ed" Oct 11 04:09:17 crc kubenswrapper[4967]: E1011 04:09:17.157542 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-neutron-agent\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ironic-neutron-agent pod=ironic-neutron-agent-657f6548d9-9fb8w_openstack(4390a950-7185-4fc0-b152-b5927940032b)\"" pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" podUID="4390a950-7185-4fc0-b152-b5927940032b" Oct 11 04:09:18 crc kubenswrapper[4967]: I1011 04:09:18.826247 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:09:18 crc kubenswrapper[4967]: I1011 04:09:18.827548 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33" containerName="ceilometer-central-agent" containerID="cri-o://61206f1f40a74f25942e30bbb070ab3e34766c2f1fe78d0ead1571eb12628643" gracePeriod=30 Oct 11 04:09:18 crc kubenswrapper[4967]: I1011 04:09:18.828302 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33" containerName="sg-core" containerID="cri-o://5e00aad7c84c286e780dce5ac301bab981272537243892e81da642ce38e7a6ce" gracePeriod=30 Oct 11 04:09:18 crc kubenswrapper[4967]: I1011 04:09:18.828325 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33" containerName="ceilometer-notification-agent" containerID="cri-o://ba24fdb99bd4eb58c090503135a2ffba357961da62cd7632202ce6f95a43418e" gracePeriod=30 Oct 11 04:09:18 crc kubenswrapper[4967]: I1011 04:09:18.828518 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33" containerName="proxy-httpd" containerID="cri-o://894b5ed10218944033f0d21b5f1591ad9b5c3e51a9d40368be249f1e22cb1c07" gracePeriod=30 Oct 11 04:09:18 crc kubenswrapper[4967]: I1011 04:09:18.832727 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 11 04:09:19 crc kubenswrapper[4967]: I1011 04:09:19.185859 4967 generic.go:334] "Generic (PLEG): container finished" podID="4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33" containerID="5e00aad7c84c286e780dce5ac301bab981272537243892e81da642ce38e7a6ce" exitCode=2 Oct 11 04:09:19 crc kubenswrapper[4967]: I1011 04:09:19.185900 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33","Type":"ContainerDied","Data":"5e00aad7c84c286e780dce5ac301bab981272537243892e81da642ce38e7a6ce"} Oct 11 04:09:19 crc kubenswrapper[4967]: I1011 04:09:19.443798 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" Oct 11 04:09:19 crc kubenswrapper[4967]: I1011 04:09:19.444464 4967 scope.go:117] "RemoveContainer" containerID="93bc6f25445cba6d1e4f70f9b3f209e2f24e716693d78b9c1971de07282202ed" Oct 11 04:09:19 crc kubenswrapper[4967]: E1011 04:09:19.444656 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-neutron-agent\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ironic-neutron-agent pod=ironic-neutron-agent-657f6548d9-9fb8w_openstack(4390a950-7185-4fc0-b152-b5927940032b)\"" pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" podUID="4390a950-7185-4fc0-b152-b5927940032b" Oct 11 04:09:20 crc kubenswrapper[4967]: I1011 04:09:20.195404 4967 generic.go:334] "Generic (PLEG): container finished" podID="4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33" containerID="894b5ed10218944033f0d21b5f1591ad9b5c3e51a9d40368be249f1e22cb1c07" exitCode=0 Oct 11 04:09:20 crc kubenswrapper[4967]: I1011 04:09:20.195435 4967 generic.go:334] "Generic (PLEG): container finished" podID="4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33" containerID="61206f1f40a74f25942e30bbb070ab3e34766c2f1fe78d0ead1571eb12628643" exitCode=0 Oct 11 04:09:20 crc kubenswrapper[4967]: I1011 04:09:20.195445 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33","Type":"ContainerDied","Data":"894b5ed10218944033f0d21b5f1591ad9b5c3e51a9d40368be249f1e22cb1c07"} Oct 11 04:09:20 crc kubenswrapper[4967]: I1011 04:09:20.195487 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33","Type":"ContainerDied","Data":"61206f1f40a74f25942e30bbb070ab3e34766c2f1fe78d0ead1571eb12628643"} Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.036364 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1117-account-create-76qc9" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.140941 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tcwld\" (UniqueName: \"kubernetes.io/projected/a758d270-0d47-415f-bae1-ae7288e90245-kube-api-access-tcwld\") pod \"a758d270-0d47-415f-bae1-ae7288e90245\" (UID: \"a758d270-0d47-415f-bae1-ae7288e90245\") " Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.142920 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.145741 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a758d270-0d47-415f-bae1-ae7288e90245-kube-api-access-tcwld" (OuterVolumeSpecName: "kube-api-access-tcwld") pod "a758d270-0d47-415f-bae1-ae7288e90245" (UID: "a758d270-0d47-415f-bae1-ae7288e90245"). InnerVolumeSpecName "kube-api-access-tcwld". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.216038 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-sync-m5k2b" event={"ID":"1d9749c8-f6a4-497f-b6dd-794865243379","Type":"ContainerStarted","Data":"12332221c5e4389a994d57ada09c21f94f33dc366ce44378ad77dceb40041f9d"} Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.220099 4967 generic.go:334] "Generic (PLEG): container finished" podID="4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33" containerID="ba24fdb99bd4eb58c090503135a2ffba357961da62cd7632202ce6f95a43418e" exitCode=0 Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.220159 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.220166 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33","Type":"ContainerDied","Data":"ba24fdb99bd4eb58c090503135a2ffba357961da62cd7632202ce6f95a43418e"} Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.220205 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33","Type":"ContainerDied","Data":"b9354f9bcc0abc860f68c16c5003b0e0deec8a4d734611fc7bcf833daa875149"} Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.220227 4967 scope.go:117] "RemoveContainer" containerID="894b5ed10218944033f0d21b5f1591ad9b5c3e51a9d40368be249f1e22cb1c07" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.224349 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-1117-account-create-76qc9" event={"ID":"a758d270-0d47-415f-bae1-ae7288e90245","Type":"ContainerDied","Data":"e02ce7ce113fc0118d9d04446289357d86af11dad8919dc24c32d14c07df80ea"} Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.224386 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="e02ce7ce113fc0118d9d04446289357d86af11dad8919dc24c32d14c07df80ea" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.224357 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-1117-account-create-76qc9" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.228486 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xh6jw" event={"ID":"d4a234ae-b210-4ac3-9bf2-841f586ee4a3","Type":"ContainerStarted","Data":"a31acefb52ee3f66f9f084ee9422d5cd5026dded84e8d4de97d3d86de40ab227"} Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.239969 4967 scope.go:117] "RemoveContainer" containerID="5e00aad7c84c286e780dce5ac301bab981272537243892e81da642ce38e7a6ce" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.243407 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnjrd\" (UniqueName: \"kubernetes.io/projected/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-kube-api-access-mnjrd\") pod \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\" (UID: \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\") " Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.243478 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-combined-ca-bundle\") pod \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\" (UID: \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\") " Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.243601 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-sg-core-conf-yaml\") pod \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\" (UID: \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\") " Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.243637 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-config-data\") pod \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\" (UID: \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\") " Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.243658 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-run-httpd\") pod \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\" (UID: \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\") " Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.243713 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-scripts\") pod \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\" (UID: \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\") " Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.243802 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-log-httpd\") pod \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\" (UID: \"4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33\") " Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.244236 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tcwld\" (UniqueName: \"kubernetes.io/projected/a758d270-0d47-415f-bae1-ae7288e90245-kube-api-access-tcwld\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.244931 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33" (UID: "4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.246138 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33" (UID: "4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.257685 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-db-sync-xh6jw" podStartSLOduration=2.207759237 podStartE2EDuration="15.257642942s" podCreationTimestamp="2025-10-11 04:09:07 +0000 UTC" firstStartedPulling="2025-10-11 04:09:08.889058777 +0000 UTC m=+1076.852267730" lastFinishedPulling="2025-10-11 04:09:21.938942492 +0000 UTC m=+1089.902151435" observedRunningTime="2025-10-11 04:09:22.254243584 +0000 UTC m=+1090.217452527" watchObservedRunningTime="2025-10-11 04:09:22.257642942 +0000 UTC m=+1090.220851885" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.261232 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-scripts" (OuterVolumeSpecName: "scripts") pod "4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33" (UID: "4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.266652 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-inspector-db-sync-m5k2b" podStartSLOduration=1.279063677 podStartE2EDuration="13.266634211s" podCreationTimestamp="2025-10-11 04:09:09 +0000 UTC" firstStartedPulling="2025-10-11 04:09:09.946032504 +0000 UTC m=+1077.909241437" lastFinishedPulling="2025-10-11 04:09:21.933603038 +0000 UTC m=+1089.896811971" observedRunningTime="2025-10-11 04:09:22.235336246 +0000 UTC m=+1090.198545179" watchObservedRunningTime="2025-10-11 04:09:22.266634211 +0000 UTC m=+1090.229843144" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.270183 4967 scope.go:117] "RemoveContainer" containerID="ba24fdb99bd4eb58c090503135a2ffba357961da62cd7632202ce6f95a43418e" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.270243 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-kube-api-access-mnjrd" (OuterVolumeSpecName: "kube-api-access-mnjrd") pod "4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33" (UID: "4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33"). InnerVolumeSpecName "kube-api-access-mnjrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.275163 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33" (UID: "4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.290655 4967 scope.go:117] "RemoveContainer" containerID="61206f1f40a74f25942e30bbb070ab3e34766c2f1fe78d0ead1571eb12628643" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.314230 4967 scope.go:117] "RemoveContainer" containerID="894b5ed10218944033f0d21b5f1591ad9b5c3e51a9d40368be249f1e22cb1c07" Oct 11 04:09:22 crc kubenswrapper[4967]: E1011 04:09:22.314634 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"894b5ed10218944033f0d21b5f1591ad9b5c3e51a9d40368be249f1e22cb1c07\": container with ID starting with 894b5ed10218944033f0d21b5f1591ad9b5c3e51a9d40368be249f1e22cb1c07 not found: ID does not exist" containerID="894b5ed10218944033f0d21b5f1591ad9b5c3e51a9d40368be249f1e22cb1c07" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.314676 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"894b5ed10218944033f0d21b5f1591ad9b5c3e51a9d40368be249f1e22cb1c07"} err="failed to get container status \"894b5ed10218944033f0d21b5f1591ad9b5c3e51a9d40368be249f1e22cb1c07\": rpc error: code = NotFound desc = could not find container \"894b5ed10218944033f0d21b5f1591ad9b5c3e51a9d40368be249f1e22cb1c07\": container with ID starting with 894b5ed10218944033f0d21b5f1591ad9b5c3e51a9d40368be249f1e22cb1c07 not found: ID does not exist" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.314700 4967 scope.go:117] "RemoveContainer" containerID="5e00aad7c84c286e780dce5ac301bab981272537243892e81da642ce38e7a6ce" Oct 11 04:09:22 crc kubenswrapper[4967]: E1011 04:09:22.314981 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5e00aad7c84c286e780dce5ac301bab981272537243892e81da642ce38e7a6ce\": container with ID starting with 5e00aad7c84c286e780dce5ac301bab981272537243892e81da642ce38e7a6ce not found: ID does not exist" containerID="5e00aad7c84c286e780dce5ac301bab981272537243892e81da642ce38e7a6ce" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.315017 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5e00aad7c84c286e780dce5ac301bab981272537243892e81da642ce38e7a6ce"} err="failed to get container status \"5e00aad7c84c286e780dce5ac301bab981272537243892e81da642ce38e7a6ce\": rpc error: code = NotFound desc = could not find container \"5e00aad7c84c286e780dce5ac301bab981272537243892e81da642ce38e7a6ce\": container with ID starting with 5e00aad7c84c286e780dce5ac301bab981272537243892e81da642ce38e7a6ce not found: ID does not exist" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.315029 4967 scope.go:117] "RemoveContainer" containerID="ba24fdb99bd4eb58c090503135a2ffba357961da62cd7632202ce6f95a43418e" Oct 11 04:09:22 crc kubenswrapper[4967]: E1011 04:09:22.315464 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ba24fdb99bd4eb58c090503135a2ffba357961da62cd7632202ce6f95a43418e\": container with ID starting with ba24fdb99bd4eb58c090503135a2ffba357961da62cd7632202ce6f95a43418e not found: ID does not exist" containerID="ba24fdb99bd4eb58c090503135a2ffba357961da62cd7632202ce6f95a43418e" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.315489 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ba24fdb99bd4eb58c090503135a2ffba357961da62cd7632202ce6f95a43418e"} err="failed to get container status \"ba24fdb99bd4eb58c090503135a2ffba357961da62cd7632202ce6f95a43418e\": rpc error: code = NotFound desc = could not find container \"ba24fdb99bd4eb58c090503135a2ffba357961da62cd7632202ce6f95a43418e\": container with ID starting with ba24fdb99bd4eb58c090503135a2ffba357961da62cd7632202ce6f95a43418e not found: ID does not exist" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.315503 4967 scope.go:117] "RemoveContainer" containerID="61206f1f40a74f25942e30bbb070ab3e34766c2f1fe78d0ead1571eb12628643" Oct 11 04:09:22 crc kubenswrapper[4967]: E1011 04:09:22.315896 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"61206f1f40a74f25942e30bbb070ab3e34766c2f1fe78d0ead1571eb12628643\": container with ID starting with 61206f1f40a74f25942e30bbb070ab3e34766c2f1fe78d0ead1571eb12628643 not found: ID does not exist" containerID="61206f1f40a74f25942e30bbb070ab3e34766c2f1fe78d0ead1571eb12628643" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.315995 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"61206f1f40a74f25942e30bbb070ab3e34766c2f1fe78d0ead1571eb12628643"} err="failed to get container status \"61206f1f40a74f25942e30bbb070ab3e34766c2f1fe78d0ead1571eb12628643\": rpc error: code = NotFound desc = could not find container \"61206f1f40a74f25942e30bbb070ab3e34766c2f1fe78d0ead1571eb12628643\": container with ID starting with 61206f1f40a74f25942e30bbb070ab3e34766c2f1fe78d0ead1571eb12628643 not found: ID does not exist" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.329928 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33" (UID: "4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.346451 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.346491 4967 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.346505 4967 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.346515 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.346526 4967 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.346537 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnjrd\" (UniqueName: \"kubernetes.io/projected/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-kube-api-access-mnjrd\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.350699 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-config-data" (OuterVolumeSpecName: "config-data") pod "4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33" (UID: "4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.448581 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.481391 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.481824 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/kube-state-metrics-0" podUID="3d120d8c-893d-49d9-8649-453997725ff2" containerName="kube-state-metrics" containerID="cri-o://4504fe990984cab7895c8f853e17d85cd0e0782bbe982534f0c7080a6b839521" gracePeriod=30 Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.581233 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.593211 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.608251 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:09:22 crc kubenswrapper[4967]: E1011 04:09:22.608725 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33" containerName="ceilometer-central-agent" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.608743 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33" containerName="ceilometer-central-agent" Oct 11 04:09:22 crc kubenswrapper[4967]: E1011 04:09:22.608769 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33" containerName="ceilometer-notification-agent" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.608777 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33" containerName="ceilometer-notification-agent" Oct 11 04:09:22 crc kubenswrapper[4967]: E1011 04:09:22.608811 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33" containerName="proxy-httpd" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.608818 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33" containerName="proxy-httpd" Oct 11 04:09:22 crc kubenswrapper[4967]: E1011 04:09:22.608832 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cb662b34-1e00-4cbb-a989-347e80fe9886" containerName="mariadb-account-create" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.608838 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="cb662b34-1e00-4cbb-a989-347e80fe9886" containerName="mariadb-account-create" Oct 11 04:09:22 crc kubenswrapper[4967]: E1011 04:09:22.608850 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33" containerName="sg-core" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.608857 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33" containerName="sg-core" Oct 11 04:09:22 crc kubenswrapper[4967]: E1011 04:09:22.608893 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a758d270-0d47-415f-bae1-ae7288e90245" containerName="mariadb-account-create" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.608899 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="a758d270-0d47-415f-bae1-ae7288e90245" containerName="mariadb-account-create" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.609119 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33" containerName="ceilometer-central-agent" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.609133 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33" containerName="sg-core" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.609174 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33" containerName="ceilometer-notification-agent" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.609184 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="a758d270-0d47-415f-bae1-ae7288e90245" containerName="mariadb-account-create" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.609195 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33" containerName="proxy-httpd" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.609204 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="cb662b34-1e00-4cbb-a989-347e80fe9886" containerName="mariadb-account-create" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.614987 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.616808 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.616841 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.623257 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.755151 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mdgjf\" (UniqueName: \"kubernetes.io/projected/2121826a-fcb0-4507-ab1e-b2309a121f2d-kube-api-access-mdgjf\") pod \"ceilometer-0\" (UID: \"2121826a-fcb0-4507-ab1e-b2309a121f2d\") " pod="openstack/ceilometer-0" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.755201 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2121826a-fcb0-4507-ab1e-b2309a121f2d-config-data\") pod \"ceilometer-0\" (UID: \"2121826a-fcb0-4507-ab1e-b2309a121f2d\") " pod="openstack/ceilometer-0" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.755227 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2121826a-fcb0-4507-ab1e-b2309a121f2d-run-httpd\") pod \"ceilometer-0\" (UID: \"2121826a-fcb0-4507-ab1e-b2309a121f2d\") " pod="openstack/ceilometer-0" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.755272 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2121826a-fcb0-4507-ab1e-b2309a121f2d-scripts\") pod \"ceilometer-0\" (UID: \"2121826a-fcb0-4507-ab1e-b2309a121f2d\") " pod="openstack/ceilometer-0" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.755316 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2121826a-fcb0-4507-ab1e-b2309a121f2d-log-httpd\") pod \"ceilometer-0\" (UID: \"2121826a-fcb0-4507-ab1e-b2309a121f2d\") " pod="openstack/ceilometer-0" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.755340 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2121826a-fcb0-4507-ab1e-b2309a121f2d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2121826a-fcb0-4507-ab1e-b2309a121f2d\") " pod="openstack/ceilometer-0" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.755373 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2121826a-fcb0-4507-ab1e-b2309a121f2d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2121826a-fcb0-4507-ab1e-b2309a121f2d\") " pod="openstack/ceilometer-0" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.826034 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33" path="/var/lib/kubelet/pods/4fa0d53b-9b58-47fb-bb38-fa4cf5b9ed33/volumes" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.869526 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2121826a-fcb0-4507-ab1e-b2309a121f2d-config-data\") pod \"ceilometer-0\" (UID: \"2121826a-fcb0-4507-ab1e-b2309a121f2d\") " pod="openstack/ceilometer-0" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.869607 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2121826a-fcb0-4507-ab1e-b2309a121f2d-run-httpd\") pod \"ceilometer-0\" (UID: \"2121826a-fcb0-4507-ab1e-b2309a121f2d\") " pod="openstack/ceilometer-0" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.869739 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2121826a-fcb0-4507-ab1e-b2309a121f2d-scripts\") pod \"ceilometer-0\" (UID: \"2121826a-fcb0-4507-ab1e-b2309a121f2d\") " pod="openstack/ceilometer-0" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.869857 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2121826a-fcb0-4507-ab1e-b2309a121f2d-log-httpd\") pod \"ceilometer-0\" (UID: \"2121826a-fcb0-4507-ab1e-b2309a121f2d\") " pod="openstack/ceilometer-0" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.869906 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2121826a-fcb0-4507-ab1e-b2309a121f2d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2121826a-fcb0-4507-ab1e-b2309a121f2d\") " pod="openstack/ceilometer-0" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.869982 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2121826a-fcb0-4507-ab1e-b2309a121f2d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2121826a-fcb0-4507-ab1e-b2309a121f2d\") " pod="openstack/ceilometer-0" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.870115 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mdgjf\" (UniqueName: \"kubernetes.io/projected/2121826a-fcb0-4507-ab1e-b2309a121f2d-kube-api-access-mdgjf\") pod \"ceilometer-0\" (UID: \"2121826a-fcb0-4507-ab1e-b2309a121f2d\") " pod="openstack/ceilometer-0" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.871796 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2121826a-fcb0-4507-ab1e-b2309a121f2d-log-httpd\") pod \"ceilometer-0\" (UID: \"2121826a-fcb0-4507-ab1e-b2309a121f2d\") " pod="openstack/ceilometer-0" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.872028 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2121826a-fcb0-4507-ab1e-b2309a121f2d-run-httpd\") pod \"ceilometer-0\" (UID: \"2121826a-fcb0-4507-ab1e-b2309a121f2d\") " pod="openstack/ceilometer-0" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.876989 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2121826a-fcb0-4507-ab1e-b2309a121f2d-scripts\") pod \"ceilometer-0\" (UID: \"2121826a-fcb0-4507-ab1e-b2309a121f2d\") " pod="openstack/ceilometer-0" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.881580 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2121826a-fcb0-4507-ab1e-b2309a121f2d-config-data\") pod \"ceilometer-0\" (UID: \"2121826a-fcb0-4507-ab1e-b2309a121f2d\") " pod="openstack/ceilometer-0" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.889172 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2121826a-fcb0-4507-ab1e-b2309a121f2d-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"2121826a-fcb0-4507-ab1e-b2309a121f2d\") " pod="openstack/ceilometer-0" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.907635 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mdgjf\" (UniqueName: \"kubernetes.io/projected/2121826a-fcb0-4507-ab1e-b2309a121f2d-kube-api-access-mdgjf\") pod \"ceilometer-0\" (UID: \"2121826a-fcb0-4507-ab1e-b2309a121f2d\") " pod="openstack/ceilometer-0" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.908489 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2121826a-fcb0-4507-ab1e-b2309a121f2d-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"2121826a-fcb0-4507-ab1e-b2309a121f2d\") " pod="openstack/ceilometer-0" Oct 11 04:09:22 crc kubenswrapper[4967]: I1011 04:09:22.957710 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.025984 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.179864 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4t756\" (UniqueName: \"kubernetes.io/projected/3d120d8c-893d-49d9-8649-453997725ff2-kube-api-access-4t756\") pod \"3d120d8c-893d-49d9-8649-453997725ff2\" (UID: \"3d120d8c-893d-49d9-8649-453997725ff2\") " Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.186332 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d120d8c-893d-49d9-8649-453997725ff2-kube-api-access-4t756" (OuterVolumeSpecName: "kube-api-access-4t756") pod "3d120d8c-893d-49d9-8649-453997725ff2" (UID: "3d120d8c-893d-49d9-8649-453997725ff2"). InnerVolumeSpecName "kube-api-access-4t756". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.245463 4967 generic.go:334] "Generic (PLEG): container finished" podID="3d120d8c-893d-49d9-8649-453997725ff2" containerID="4504fe990984cab7895c8f853e17d85cd0e0782bbe982534f0c7080a6b839521" exitCode=2 Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.245525 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3d120d8c-893d-49d9-8649-453997725ff2","Type":"ContainerDied","Data":"4504fe990984cab7895c8f853e17d85cd0e0782bbe982534f0c7080a6b839521"} Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.245552 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.245574 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"3d120d8c-893d-49d9-8649-453997725ff2","Type":"ContainerDied","Data":"18078965abda972022268296e54fa6994f5e9c47809a5360b54b182a2cc6ceb1"} Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.245592 4967 scope.go:117] "RemoveContainer" containerID="4504fe990984cab7895c8f853e17d85cd0e0782bbe982534f0c7080a6b839521" Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.282827 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4t756\" (UniqueName: \"kubernetes.io/projected/3d120d8c-893d-49d9-8649-453997725ff2-kube-api-access-4t756\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.284102 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.287343 4967 scope.go:117] "RemoveContainer" containerID="4504fe990984cab7895c8f853e17d85cd0e0782bbe982534f0c7080a6b839521" Oct 11 04:09:23 crc kubenswrapper[4967]: E1011 04:09:23.292977 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4504fe990984cab7895c8f853e17d85cd0e0782bbe982534f0c7080a6b839521\": container with ID starting with 4504fe990984cab7895c8f853e17d85cd0e0782bbe982534f0c7080a6b839521 not found: ID does not exist" containerID="4504fe990984cab7895c8f853e17d85cd0e0782bbe982534f0c7080a6b839521" Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.293025 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4504fe990984cab7895c8f853e17d85cd0e0782bbe982534f0c7080a6b839521"} err="failed to get container status \"4504fe990984cab7895c8f853e17d85cd0e0782bbe982534f0c7080a6b839521\": rpc error: code = NotFound desc = could not find container \"4504fe990984cab7895c8f853e17d85cd0e0782bbe982534f0c7080a6b839521\": container with ID starting with 4504fe990984cab7895c8f853e17d85cd0e0782bbe982534f0c7080a6b839521 not found: ID does not exist" Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.298989 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.313200 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/kube-state-metrics-0"] Oct 11 04:09:23 crc kubenswrapper[4967]: E1011 04:09:23.313784 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d120d8c-893d-49d9-8649-453997725ff2" containerName="kube-state-metrics" Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.313821 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d120d8c-893d-49d9-8649-453997725ff2" containerName="kube-state-metrics" Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.314055 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d120d8c-893d-49d9-8649-453997725ff2" containerName="kube-state-metrics" Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.315429 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.318764 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-kube-state-metrics-svc" Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.319001 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"kube-state-metrics-tls-config" Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.321742 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.486467 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5917a28-16f9-4b7b-8b12-04c44c5333cf-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"f5917a28-16f9-4b7b-8b12-04c44c5333cf\") " pod="openstack/kube-state-metrics-0" Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.486867 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5917a28-16f9-4b7b-8b12-04c44c5333cf-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"f5917a28-16f9-4b7b-8b12-04c44c5333cf\") " pod="openstack/kube-state-metrics-0" Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.487327 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/f5917a28-16f9-4b7b-8b12-04c44c5333cf-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"f5917a28-16f9-4b7b-8b12-04c44c5333cf\") " pod="openstack/kube-state-metrics-0" Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.487387 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mz2wl\" (UniqueName: \"kubernetes.io/projected/f5917a28-16f9-4b7b-8b12-04c44c5333cf-kube-api-access-mz2wl\") pod \"kube-state-metrics-0\" (UID: \"f5917a28-16f9-4b7b-8b12-04c44c5333cf\") " pod="openstack/kube-state-metrics-0" Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.490250 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.589067 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5917a28-16f9-4b7b-8b12-04c44c5333cf-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"f5917a28-16f9-4b7b-8b12-04c44c5333cf\") " pod="openstack/kube-state-metrics-0" Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.589755 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/f5917a28-16f9-4b7b-8b12-04c44c5333cf-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"f5917a28-16f9-4b7b-8b12-04c44c5333cf\") " pod="openstack/kube-state-metrics-0" Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.589789 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mz2wl\" (UniqueName: \"kubernetes.io/projected/f5917a28-16f9-4b7b-8b12-04c44c5333cf-kube-api-access-mz2wl\") pod \"kube-state-metrics-0\" (UID: \"f5917a28-16f9-4b7b-8b12-04c44c5333cf\") " pod="openstack/kube-state-metrics-0" Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.589819 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5917a28-16f9-4b7b-8b12-04c44c5333cf-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"f5917a28-16f9-4b7b-8b12-04c44c5333cf\") " pod="openstack/kube-state-metrics-0" Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.593591 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-config\" (UniqueName: \"kubernetes.io/secret/f5917a28-16f9-4b7b-8b12-04c44c5333cf-kube-state-metrics-tls-config\") pod \"kube-state-metrics-0\" (UID: \"f5917a28-16f9-4b7b-8b12-04c44c5333cf\") " pod="openstack/kube-state-metrics-0" Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.594296 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/f5917a28-16f9-4b7b-8b12-04c44c5333cf-combined-ca-bundle\") pod \"kube-state-metrics-0\" (UID: \"f5917a28-16f9-4b7b-8b12-04c44c5333cf\") " pod="openstack/kube-state-metrics-0" Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.595331 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-state-metrics-tls-certs\" (UniqueName: \"kubernetes.io/secret/f5917a28-16f9-4b7b-8b12-04c44c5333cf-kube-state-metrics-tls-certs\") pod \"kube-state-metrics-0\" (UID: \"f5917a28-16f9-4b7b-8b12-04c44c5333cf\") " pod="openstack/kube-state-metrics-0" Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.614813 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mz2wl\" (UniqueName: \"kubernetes.io/projected/f5917a28-16f9-4b7b-8b12-04c44c5333cf-kube-api-access-mz2wl\") pod \"kube-state-metrics-0\" (UID: \"f5917a28-16f9-4b7b-8b12-04c44c5333cf\") " pod="openstack/kube-state-metrics-0" Oct 11 04:09:23 crc kubenswrapper[4967]: I1011 04:09:23.647376 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/kube-state-metrics-0" Oct 11 04:09:24 crc kubenswrapper[4967]: I1011 04:09:24.145139 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/kube-state-metrics-0"] Oct 11 04:09:24 crc kubenswrapper[4967]: W1011 04:09:24.160738 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf5917a28_16f9_4b7b_8b12_04c44c5333cf.slice/crio-fe8e3801eaf41f65e201b3f5f680f007bd007bb746fdcae6edfc523b0cdd15c7 WatchSource:0}: Error finding container fe8e3801eaf41f65e201b3f5f680f007bd007bb746fdcae6edfc523b0cdd15c7: Status 404 returned error can't find the container with id fe8e3801eaf41f65e201b3f5f680f007bd007bb746fdcae6edfc523b0cdd15c7 Oct 11 04:09:24 crc kubenswrapper[4967]: I1011 04:09:24.261547 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f5917a28-16f9-4b7b-8b12-04c44c5333cf","Type":"ContainerStarted","Data":"fe8e3801eaf41f65e201b3f5f680f007bd007bb746fdcae6edfc523b0cdd15c7"} Oct 11 04:09:24 crc kubenswrapper[4967]: I1011 04:09:24.266135 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2121826a-fcb0-4507-ab1e-b2309a121f2d","Type":"ContainerStarted","Data":"a756d89f1be34c7ba6fd0c7fb49236ce3ca75479a44de0b105b460c60d9a787e"} Oct 11 04:09:24 crc kubenswrapper[4967]: I1011 04:09:24.829113 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d120d8c-893d-49d9-8649-453997725ff2" path="/var/lib/kubelet/pods/3d120d8c-893d-49d9-8649-453997725ff2/volumes" Oct 11 04:09:24 crc kubenswrapper[4967]: I1011 04:09:24.982975 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:09:25 crc kubenswrapper[4967]: I1011 04:09:25.282484 4967 generic.go:334] "Generic (PLEG): container finished" podID="1d9749c8-f6a4-497f-b6dd-794865243379" containerID="12332221c5e4389a994d57ada09c21f94f33dc366ce44378ad77dceb40041f9d" exitCode=0 Oct 11 04:09:25 crc kubenswrapper[4967]: I1011 04:09:25.282546 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-sync-m5k2b" event={"ID":"1d9749c8-f6a4-497f-b6dd-794865243379","Type":"ContainerDied","Data":"12332221c5e4389a994d57ada09c21f94f33dc366ce44378ad77dceb40041f9d"} Oct 11 04:09:25 crc kubenswrapper[4967]: I1011 04:09:25.285452 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/kube-state-metrics-0" event={"ID":"f5917a28-16f9-4b7b-8b12-04c44c5333cf","Type":"ContainerStarted","Data":"d1aaa0dd52add55f39239cfb5617d50f5ec092cecd4e4092aa9af8f095a17565"} Oct 11 04:09:25 crc kubenswrapper[4967]: I1011 04:09:25.285621 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/kube-state-metrics-0" Oct 11 04:09:25 crc kubenswrapper[4967]: I1011 04:09:25.288250 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2121826a-fcb0-4507-ab1e-b2309a121f2d","Type":"ContainerStarted","Data":"2245b260f9b40946fb319eef185b4333c58bf0dd4e6b0b9fdcd6d0455e30ea29"} Oct 11 04:09:25 crc kubenswrapper[4967]: I1011 04:09:25.288279 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2121826a-fcb0-4507-ab1e-b2309a121f2d","Type":"ContainerStarted","Data":"e653793c0d8eee960244df6319727358785d02b58f4017b8f097cd5901bad763"} Oct 11 04:09:25 crc kubenswrapper[4967]: I1011 04:09:25.328233 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/kube-state-metrics-0" podStartSLOduration=1.900727259 podStartE2EDuration="2.32821677s" podCreationTimestamp="2025-10-11 04:09:23 +0000 UTC" firstStartedPulling="2025-10-11 04:09:24.163447816 +0000 UTC m=+1092.126656789" lastFinishedPulling="2025-10-11 04:09:24.590937367 +0000 UTC m=+1092.554146300" observedRunningTime="2025-10-11 04:09:25.321932225 +0000 UTC m=+1093.285141158" watchObservedRunningTime="2025-10-11 04:09:25.32821677 +0000 UTC m=+1093.291425693" Oct 11 04:09:26 crc kubenswrapper[4967]: I1011 04:09:26.298994 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2121826a-fcb0-4507-ab1e-b2309a121f2d","Type":"ContainerStarted","Data":"9223804340b64318a1b6a32161e179568834ac90a98ab57c8439d20bbd0969cd"} Oct 11 04:09:26 crc kubenswrapper[4967]: I1011 04:09:26.626920 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-sync-m5k2b" Oct 11 04:09:26 crc kubenswrapper[4967]: I1011 04:09:26.760167 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d9749c8-f6a4-497f-b6dd-794865243379-scripts\") pod \"1d9749c8-f6a4-497f-b6dd-794865243379\" (UID: \"1d9749c8-f6a4-497f-b6dd-794865243379\") " Oct 11 04:09:26 crc kubenswrapper[4967]: I1011 04:09:26.760435 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/1d9749c8-f6a4-497f-b6dd-794865243379-var-lib-ironic\") pod \"1d9749c8-f6a4-497f-b6dd-794865243379\" (UID: \"1d9749c8-f6a4-497f-b6dd-794865243379\") " Oct 11 04:09:26 crc kubenswrapper[4967]: I1011 04:09:26.760539 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d9749c8-f6a4-497f-b6dd-794865243379-combined-ca-bundle\") pod \"1d9749c8-f6a4-497f-b6dd-794865243379\" (UID: \"1d9749c8-f6a4-497f-b6dd-794865243379\") " Oct 11 04:09:26 crc kubenswrapper[4967]: I1011 04:09:26.760561 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/1d9749c8-f6a4-497f-b6dd-794865243379-config\") pod \"1d9749c8-f6a4-497f-b6dd-794865243379\" (UID: \"1d9749c8-f6a4-497f-b6dd-794865243379\") " Oct 11 04:09:26 crc kubenswrapper[4967]: I1011 04:09:26.760693 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qj465\" (UniqueName: \"kubernetes.io/projected/1d9749c8-f6a4-497f-b6dd-794865243379-kube-api-access-qj465\") pod \"1d9749c8-f6a4-497f-b6dd-794865243379\" (UID: \"1d9749c8-f6a4-497f-b6dd-794865243379\") " Oct 11 04:09:26 crc kubenswrapper[4967]: I1011 04:09:26.760852 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/1d9749c8-f6a4-497f-b6dd-794865243379-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"1d9749c8-f6a4-497f-b6dd-794865243379\" (UID: \"1d9749c8-f6a4-497f-b6dd-794865243379\") " Oct 11 04:09:26 crc kubenswrapper[4967]: I1011 04:09:26.760934 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/1d9749c8-f6a4-497f-b6dd-794865243379-etc-podinfo\") pod \"1d9749c8-f6a4-497f-b6dd-794865243379\" (UID: \"1d9749c8-f6a4-497f-b6dd-794865243379\") " Oct 11 04:09:26 crc kubenswrapper[4967]: I1011 04:09:26.768550 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d9749c8-f6a4-497f-b6dd-794865243379-var-lib-ironic" (OuterVolumeSpecName: "var-lib-ironic") pod "1d9749c8-f6a4-497f-b6dd-794865243379" (UID: "1d9749c8-f6a4-497f-b6dd-794865243379"). InnerVolumeSpecName "var-lib-ironic". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:09:26 crc kubenswrapper[4967]: I1011 04:09:26.771256 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d9749c8-f6a4-497f-b6dd-794865243379-var-lib-ironic-inspector-dhcp-hostsdir" (OuterVolumeSpecName: "var-lib-ironic-inspector-dhcp-hostsdir") pod "1d9749c8-f6a4-497f-b6dd-794865243379" (UID: "1d9749c8-f6a4-497f-b6dd-794865243379"). InnerVolumeSpecName "var-lib-ironic-inspector-dhcp-hostsdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:09:26 crc kubenswrapper[4967]: I1011 04:09:26.775368 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/1d9749c8-f6a4-497f-b6dd-794865243379-etc-podinfo" (OuterVolumeSpecName: "etc-podinfo") pod "1d9749c8-f6a4-497f-b6dd-794865243379" (UID: "1d9749c8-f6a4-497f-b6dd-794865243379"). InnerVolumeSpecName "etc-podinfo". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 11 04:09:26 crc kubenswrapper[4967]: I1011 04:09:26.776135 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d9749c8-f6a4-497f-b6dd-794865243379-scripts" (OuterVolumeSpecName: "scripts") pod "1d9749c8-f6a4-497f-b6dd-794865243379" (UID: "1d9749c8-f6a4-497f-b6dd-794865243379"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:09:26 crc kubenswrapper[4967]: I1011 04:09:26.785472 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d9749c8-f6a4-497f-b6dd-794865243379-kube-api-access-qj465" (OuterVolumeSpecName: "kube-api-access-qj465") pod "1d9749c8-f6a4-497f-b6dd-794865243379" (UID: "1d9749c8-f6a4-497f-b6dd-794865243379"). InnerVolumeSpecName "kube-api-access-qj465". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:09:26 crc kubenswrapper[4967]: I1011 04:09:26.789624 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d9749c8-f6a4-497f-b6dd-794865243379-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1d9749c8-f6a4-497f-b6dd-794865243379" (UID: "1d9749c8-f6a4-497f-b6dd-794865243379"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:09:26 crc kubenswrapper[4967]: I1011 04:09:26.798515 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1d9749c8-f6a4-497f-b6dd-794865243379-config" (OuterVolumeSpecName: "config") pod "1d9749c8-f6a4-497f-b6dd-794865243379" (UID: "1d9749c8-f6a4-497f-b6dd-794865243379"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:09:26 crc kubenswrapper[4967]: I1011 04:09:26.863563 4967 reconciler_common.go:293] "Volume detached for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/1d9749c8-f6a4-497f-b6dd-794865243379-var-lib-ironic\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:26 crc kubenswrapper[4967]: I1011 04:09:26.863810 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1d9749c8-f6a4-497f-b6dd-794865243379-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:26 crc kubenswrapper[4967]: I1011 04:09:26.863821 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/1d9749c8-f6a4-497f-b6dd-794865243379-config\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:26 crc kubenswrapper[4967]: I1011 04:09:26.863831 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qj465\" (UniqueName: \"kubernetes.io/projected/1d9749c8-f6a4-497f-b6dd-794865243379-kube-api-access-qj465\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:26 crc kubenswrapper[4967]: I1011 04:09:26.863843 4967 reconciler_common.go:293] "Volume detached for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/1d9749c8-f6a4-497f-b6dd-794865243379-var-lib-ironic-inspector-dhcp-hostsdir\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:26 crc kubenswrapper[4967]: I1011 04:09:26.863851 4967 reconciler_common.go:293] "Volume detached for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/1d9749c8-f6a4-497f-b6dd-794865243379-etc-podinfo\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:26 crc kubenswrapper[4967]: I1011 04:09:26.863861 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/1d9749c8-f6a4-497f-b6dd-794865243379-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:27 crc kubenswrapper[4967]: I1011 04:09:27.311481 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-db-sync-m5k2b" event={"ID":"1d9749c8-f6a4-497f-b6dd-794865243379","Type":"ContainerDied","Data":"19e8690caa331f8312237a061989f958f1bd1b07c3a4204df852d5e5014144f3"} Oct 11 04:09:27 crc kubenswrapper[4967]: I1011 04:09:27.311546 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="19e8690caa331f8312237a061989f958f1bd1b07c3a4204df852d5e5014144f3" Oct 11 04:09:27 crc kubenswrapper[4967]: I1011 04:09:27.311611 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-db-sync-m5k2b" Oct 11 04:09:28 crc kubenswrapper[4967]: I1011 04:09:28.324586 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2121826a-fcb0-4507-ab1e-b2309a121f2d","Type":"ContainerStarted","Data":"7737db67bd8c3a72bfb6f29bfeb0b66eaeeb58b98c86d55c0440da7faf3e89ca"} Oct 11 04:09:28 crc kubenswrapper[4967]: I1011 04:09:28.324911 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 11 04:09:28 crc kubenswrapper[4967]: I1011 04:09:28.324790 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2121826a-fcb0-4507-ab1e-b2309a121f2d" containerName="ceilometer-central-agent" containerID="cri-o://e653793c0d8eee960244df6319727358785d02b58f4017b8f097cd5901bad763" gracePeriod=30 Oct 11 04:09:28 crc kubenswrapper[4967]: I1011 04:09:28.324975 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2121826a-fcb0-4507-ab1e-b2309a121f2d" containerName="sg-core" containerID="cri-o://9223804340b64318a1b6a32161e179568834ac90a98ab57c8439d20bbd0969cd" gracePeriod=30 Oct 11 04:09:28 crc kubenswrapper[4967]: I1011 04:09:28.325032 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2121826a-fcb0-4507-ab1e-b2309a121f2d" containerName="proxy-httpd" containerID="cri-o://7737db67bd8c3a72bfb6f29bfeb0b66eaeeb58b98c86d55c0440da7faf3e89ca" gracePeriod=30 Oct 11 04:09:28 crc kubenswrapper[4967]: I1011 04:09:28.325066 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="2121826a-fcb0-4507-ab1e-b2309a121f2d" containerName="ceilometer-notification-agent" containerID="cri-o://2245b260f9b40946fb319eef185b4333c58bf0dd4e6b0b9fdcd6d0455e30ea29" gracePeriod=30 Oct 11 04:09:28 crc kubenswrapper[4967]: I1011 04:09:28.354286 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.417374057 podStartE2EDuration="6.354271256s" podCreationTimestamp="2025-10-11 04:09:22 +0000 UTC" firstStartedPulling="2025-10-11 04:09:23.501949688 +0000 UTC m=+1091.465158621" lastFinishedPulling="2025-10-11 04:09:27.438846887 +0000 UTC m=+1095.402055820" observedRunningTime="2025-10-11 04:09:28.35356341 +0000 UTC m=+1096.316772353" watchObservedRunningTime="2025-10-11 04:09:28.354271256 +0000 UTC m=+1096.317480179" Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.336039 4967 generic.go:334] "Generic (PLEG): container finished" podID="2121826a-fcb0-4507-ab1e-b2309a121f2d" containerID="7737db67bd8c3a72bfb6f29bfeb0b66eaeeb58b98c86d55c0440da7faf3e89ca" exitCode=0 Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.336368 4967 generic.go:334] "Generic (PLEG): container finished" podID="2121826a-fcb0-4507-ab1e-b2309a121f2d" containerID="9223804340b64318a1b6a32161e179568834ac90a98ab57c8439d20bbd0969cd" exitCode=2 Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.336379 4967 generic.go:334] "Generic (PLEG): container finished" podID="2121826a-fcb0-4507-ab1e-b2309a121f2d" containerID="2245b260f9b40946fb319eef185b4333c58bf0dd4e6b0b9fdcd6d0455e30ea29" exitCode=0 Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.336110 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2121826a-fcb0-4507-ab1e-b2309a121f2d","Type":"ContainerDied","Data":"7737db67bd8c3a72bfb6f29bfeb0b66eaeeb58b98c86d55c0440da7faf3e89ca"} Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.336415 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2121826a-fcb0-4507-ab1e-b2309a121f2d","Type":"ContainerDied","Data":"9223804340b64318a1b6a32161e179568834ac90a98ab57c8439d20bbd0969cd"} Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.336433 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2121826a-fcb0-4507-ab1e-b2309a121f2d","Type":"ContainerDied","Data":"2245b260f9b40946fb319eef185b4333c58bf0dd4e6b0b9fdcd6d0455e30ea29"} Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.365979 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-inspector-0"] Oct 11 04:09:29 crc kubenswrapper[4967]: E1011 04:09:29.366461 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1d9749c8-f6a4-497f-b6dd-794865243379" containerName="ironic-inspector-db-sync" Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.366479 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="1d9749c8-f6a4-497f-b6dd-794865243379" containerName="ironic-inspector-db-sync" Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.366721 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="1d9749c8-f6a4-497f-b6dd-794865243379" containerName="ironic-inspector-db-sync" Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.369812 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-0" Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.372998 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-config-data" Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.373151 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-scripts" Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.379809 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-0"] Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.514718 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/95c0b2ec-d176-4832-897b-7c33176243cd-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-0\" (UID: \"95c0b2ec-d176-4832-897b-7c33176243cd\") " pod="openstack/ironic-inspector-0" Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.514765 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/95c0b2ec-d176-4832-897b-7c33176243cd-etc-podinfo\") pod \"ironic-inspector-0\" (UID: \"95c0b2ec-d176-4832-897b-7c33176243cd\") " pod="openstack/ironic-inspector-0" Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.514789 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95c0b2ec-d176-4832-897b-7c33176243cd-combined-ca-bundle\") pod \"ironic-inspector-0\" (UID: \"95c0b2ec-d176-4832-897b-7c33176243cd\") " pod="openstack/ironic-inspector-0" Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.514855 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/95c0b2ec-d176-4832-897b-7c33176243cd-config\") pod \"ironic-inspector-0\" (UID: \"95c0b2ec-d176-4832-897b-7c33176243cd\") " pod="openstack/ironic-inspector-0" Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.514886 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/95c0b2ec-d176-4832-897b-7c33176243cd-var-lib-ironic\") pod \"ironic-inspector-0\" (UID: \"95c0b2ec-d176-4832-897b-7c33176243cd\") " pod="openstack/ironic-inspector-0" Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.514936 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rbmng\" (UniqueName: \"kubernetes.io/projected/95c0b2ec-d176-4832-897b-7c33176243cd-kube-api-access-rbmng\") pod \"ironic-inspector-0\" (UID: \"95c0b2ec-d176-4832-897b-7c33176243cd\") " pod="openstack/ironic-inspector-0" Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.514961 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95c0b2ec-d176-4832-897b-7c33176243cd-scripts\") pod \"ironic-inspector-0\" (UID: \"95c0b2ec-d176-4832-897b-7c33176243cd\") " pod="openstack/ironic-inspector-0" Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.616679 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95c0b2ec-d176-4832-897b-7c33176243cd-combined-ca-bundle\") pod \"ironic-inspector-0\" (UID: \"95c0b2ec-d176-4832-897b-7c33176243cd\") " pod="openstack/ironic-inspector-0" Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.616779 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/95c0b2ec-d176-4832-897b-7c33176243cd-config\") pod \"ironic-inspector-0\" (UID: \"95c0b2ec-d176-4832-897b-7c33176243cd\") " pod="openstack/ironic-inspector-0" Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.616822 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/95c0b2ec-d176-4832-897b-7c33176243cd-var-lib-ironic\") pod \"ironic-inspector-0\" (UID: \"95c0b2ec-d176-4832-897b-7c33176243cd\") " pod="openstack/ironic-inspector-0" Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.616875 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rbmng\" (UniqueName: \"kubernetes.io/projected/95c0b2ec-d176-4832-897b-7c33176243cd-kube-api-access-rbmng\") pod \"ironic-inspector-0\" (UID: \"95c0b2ec-d176-4832-897b-7c33176243cd\") " pod="openstack/ironic-inspector-0" Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.616900 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95c0b2ec-d176-4832-897b-7c33176243cd-scripts\") pod \"ironic-inspector-0\" (UID: \"95c0b2ec-d176-4832-897b-7c33176243cd\") " pod="openstack/ironic-inspector-0" Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.616929 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/95c0b2ec-d176-4832-897b-7c33176243cd-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-0\" (UID: \"95c0b2ec-d176-4832-897b-7c33176243cd\") " pod="openstack/ironic-inspector-0" Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.616946 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/95c0b2ec-d176-4832-897b-7c33176243cd-etc-podinfo\") pod \"ironic-inspector-0\" (UID: \"95c0b2ec-d176-4832-897b-7c33176243cd\") " pod="openstack/ironic-inspector-0" Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.618122 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/95c0b2ec-d176-4832-897b-7c33176243cd-var-lib-ironic\") pod \"ironic-inspector-0\" (UID: \"95c0b2ec-d176-4832-897b-7c33176243cd\") " pod="openstack/ironic-inspector-0" Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.618886 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/95c0b2ec-d176-4832-897b-7c33176243cd-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-0\" (UID: \"95c0b2ec-d176-4832-897b-7c33176243cd\") " pod="openstack/ironic-inspector-0" Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.623719 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/95c0b2ec-d176-4832-897b-7c33176243cd-etc-podinfo\") pod \"ironic-inspector-0\" (UID: \"95c0b2ec-d176-4832-897b-7c33176243cd\") " pod="openstack/ironic-inspector-0" Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.625286 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95c0b2ec-d176-4832-897b-7c33176243cd-combined-ca-bundle\") pod \"ironic-inspector-0\" (UID: \"95c0b2ec-d176-4832-897b-7c33176243cd\") " pod="openstack/ironic-inspector-0" Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.625475 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95c0b2ec-d176-4832-897b-7c33176243cd-scripts\") pod \"ironic-inspector-0\" (UID: \"95c0b2ec-d176-4832-897b-7c33176243cd\") " pod="openstack/ironic-inspector-0" Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.627971 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/95c0b2ec-d176-4832-897b-7c33176243cd-config\") pod \"ironic-inspector-0\" (UID: \"95c0b2ec-d176-4832-897b-7c33176243cd\") " pod="openstack/ironic-inspector-0" Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.639045 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rbmng\" (UniqueName: \"kubernetes.io/projected/95c0b2ec-d176-4832-897b-7c33176243cd-kube-api-access-rbmng\") pod \"ironic-inspector-0\" (UID: \"95c0b2ec-d176-4832-897b-7c33176243cd\") " pod="openstack/ironic-inspector-0" Oct 11 04:09:29 crc kubenswrapper[4967]: I1011 04:09:29.693263 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-0" Oct 11 04:09:30 crc kubenswrapper[4967]: I1011 04:09:30.142665 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-0"] Oct 11 04:09:30 crc kubenswrapper[4967]: I1011 04:09:30.345338 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"95c0b2ec-d176-4832-897b-7c33176243cd","Type":"ContainerStarted","Data":"cb8b484056c96f18473c916e31fbf8c80627fb448096b792c38c4ef2b2bb6e72"} Oct 11 04:09:31 crc kubenswrapper[4967]: I1011 04:09:31.274658 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-inspector-0"] Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.227302 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.378087 4967 generic.go:334] "Generic (PLEG): container finished" podID="2121826a-fcb0-4507-ab1e-b2309a121f2d" containerID="e653793c0d8eee960244df6319727358785d02b58f4017b8f097cd5901bad763" exitCode=0 Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.378190 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2121826a-fcb0-4507-ab1e-b2309a121f2d","Type":"ContainerDied","Data":"e653793c0d8eee960244df6319727358785d02b58f4017b8f097cd5901bad763"} Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.378222 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"2121826a-fcb0-4507-ab1e-b2309a121f2d","Type":"ContainerDied","Data":"a756d89f1be34c7ba6fd0c7fb49236ce3ca75479a44de0b105b460c60d9a787e"} Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.378242 4967 scope.go:117] "RemoveContainer" containerID="7737db67bd8c3a72bfb6f29bfeb0b66eaeeb58b98c86d55c0440da7faf3e89ca" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.378392 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.407835 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2121826a-fcb0-4507-ab1e-b2309a121f2d-scripts\") pod \"2121826a-fcb0-4507-ab1e-b2309a121f2d\" (UID: \"2121826a-fcb0-4507-ab1e-b2309a121f2d\") " Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.407964 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2121826a-fcb0-4507-ab1e-b2309a121f2d-log-httpd\") pod \"2121826a-fcb0-4507-ab1e-b2309a121f2d\" (UID: \"2121826a-fcb0-4507-ab1e-b2309a121f2d\") " Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.408009 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2121826a-fcb0-4507-ab1e-b2309a121f2d-sg-core-conf-yaml\") pod \"2121826a-fcb0-4507-ab1e-b2309a121f2d\" (UID: \"2121826a-fcb0-4507-ab1e-b2309a121f2d\") " Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.408130 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2121826a-fcb0-4507-ab1e-b2309a121f2d-config-data\") pod \"2121826a-fcb0-4507-ab1e-b2309a121f2d\" (UID: \"2121826a-fcb0-4507-ab1e-b2309a121f2d\") " Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.408164 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2121826a-fcb0-4507-ab1e-b2309a121f2d-combined-ca-bundle\") pod \"2121826a-fcb0-4507-ab1e-b2309a121f2d\" (UID: \"2121826a-fcb0-4507-ab1e-b2309a121f2d\") " Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.408208 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mdgjf\" (UniqueName: \"kubernetes.io/projected/2121826a-fcb0-4507-ab1e-b2309a121f2d-kube-api-access-mdgjf\") pod \"2121826a-fcb0-4507-ab1e-b2309a121f2d\" (UID: \"2121826a-fcb0-4507-ab1e-b2309a121f2d\") " Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.408374 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2121826a-fcb0-4507-ab1e-b2309a121f2d-run-httpd\") pod \"2121826a-fcb0-4507-ab1e-b2309a121f2d\" (UID: \"2121826a-fcb0-4507-ab1e-b2309a121f2d\") " Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.409833 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2121826a-fcb0-4507-ab1e-b2309a121f2d-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "2121826a-fcb0-4507-ab1e-b2309a121f2d" (UID: "2121826a-fcb0-4507-ab1e-b2309a121f2d"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.409979 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2121826a-fcb0-4507-ab1e-b2309a121f2d-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "2121826a-fcb0-4507-ab1e-b2309a121f2d" (UID: "2121826a-fcb0-4507-ab1e-b2309a121f2d"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.413455 4967 scope.go:117] "RemoveContainer" containerID="9223804340b64318a1b6a32161e179568834ac90a98ab57c8439d20bbd0969cd" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.415134 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2121826a-fcb0-4507-ab1e-b2309a121f2d-kube-api-access-mdgjf" (OuterVolumeSpecName: "kube-api-access-mdgjf") pod "2121826a-fcb0-4507-ab1e-b2309a121f2d" (UID: "2121826a-fcb0-4507-ab1e-b2309a121f2d"). InnerVolumeSpecName "kube-api-access-mdgjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.432517 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2121826a-fcb0-4507-ab1e-b2309a121f2d-scripts" (OuterVolumeSpecName: "scripts") pod "2121826a-fcb0-4507-ab1e-b2309a121f2d" (UID: "2121826a-fcb0-4507-ab1e-b2309a121f2d"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.440414 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2121826a-fcb0-4507-ab1e-b2309a121f2d-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "2121826a-fcb0-4507-ab1e-b2309a121f2d" (UID: "2121826a-fcb0-4507-ab1e-b2309a121f2d"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.492736 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2121826a-fcb0-4507-ab1e-b2309a121f2d-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "2121826a-fcb0-4507-ab1e-b2309a121f2d" (UID: "2121826a-fcb0-4507-ab1e-b2309a121f2d"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.510552 4967 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/2121826a-fcb0-4507-ab1e-b2309a121f2d-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.510865 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2121826a-fcb0-4507-ab1e-b2309a121f2d-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.510978 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mdgjf\" (UniqueName: \"kubernetes.io/projected/2121826a-fcb0-4507-ab1e-b2309a121f2d-kube-api-access-mdgjf\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.511000 4967 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2121826a-fcb0-4507-ab1e-b2309a121f2d-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.511012 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/2121826a-fcb0-4507-ab1e-b2309a121f2d-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.511023 4967 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/2121826a-fcb0-4507-ab1e-b2309a121f2d-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.525965 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/2121826a-fcb0-4507-ab1e-b2309a121f2d-config-data" (OuterVolumeSpecName: "config-data") pod "2121826a-fcb0-4507-ab1e-b2309a121f2d" (UID: "2121826a-fcb0-4507-ab1e-b2309a121f2d"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.612866 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2121826a-fcb0-4507-ab1e-b2309a121f2d-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.617943 4967 scope.go:117] "RemoveContainer" containerID="2245b260f9b40946fb319eef185b4333c58bf0dd4e6b0b9fdcd6d0455e30ea29" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.645683 4967 scope.go:117] "RemoveContainer" containerID="e653793c0d8eee960244df6319727358785d02b58f4017b8f097cd5901bad763" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.663170 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/kube-state-metrics-0" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.671403 4967 scope.go:117] "RemoveContainer" containerID="7737db67bd8c3a72bfb6f29bfeb0b66eaeeb58b98c86d55c0440da7faf3e89ca" Oct 11 04:09:33 crc kubenswrapper[4967]: E1011 04:09:33.672041 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7737db67bd8c3a72bfb6f29bfeb0b66eaeeb58b98c86d55c0440da7faf3e89ca\": container with ID starting with 7737db67bd8c3a72bfb6f29bfeb0b66eaeeb58b98c86d55c0440da7faf3e89ca not found: ID does not exist" containerID="7737db67bd8c3a72bfb6f29bfeb0b66eaeeb58b98c86d55c0440da7faf3e89ca" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.672134 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7737db67bd8c3a72bfb6f29bfeb0b66eaeeb58b98c86d55c0440da7faf3e89ca"} err="failed to get container status \"7737db67bd8c3a72bfb6f29bfeb0b66eaeeb58b98c86d55c0440da7faf3e89ca\": rpc error: code = NotFound desc = could not find container \"7737db67bd8c3a72bfb6f29bfeb0b66eaeeb58b98c86d55c0440da7faf3e89ca\": container with ID starting with 7737db67bd8c3a72bfb6f29bfeb0b66eaeeb58b98c86d55c0440da7faf3e89ca not found: ID does not exist" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.672159 4967 scope.go:117] "RemoveContainer" containerID="9223804340b64318a1b6a32161e179568834ac90a98ab57c8439d20bbd0969cd" Oct 11 04:09:33 crc kubenswrapper[4967]: E1011 04:09:33.672504 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9223804340b64318a1b6a32161e179568834ac90a98ab57c8439d20bbd0969cd\": container with ID starting with 9223804340b64318a1b6a32161e179568834ac90a98ab57c8439d20bbd0969cd not found: ID does not exist" containerID="9223804340b64318a1b6a32161e179568834ac90a98ab57c8439d20bbd0969cd" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.672553 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9223804340b64318a1b6a32161e179568834ac90a98ab57c8439d20bbd0969cd"} err="failed to get container status \"9223804340b64318a1b6a32161e179568834ac90a98ab57c8439d20bbd0969cd\": rpc error: code = NotFound desc = could not find container \"9223804340b64318a1b6a32161e179568834ac90a98ab57c8439d20bbd0969cd\": container with ID starting with 9223804340b64318a1b6a32161e179568834ac90a98ab57c8439d20bbd0969cd not found: ID does not exist" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.672587 4967 scope.go:117] "RemoveContainer" containerID="2245b260f9b40946fb319eef185b4333c58bf0dd4e6b0b9fdcd6d0455e30ea29" Oct 11 04:09:33 crc kubenswrapper[4967]: E1011 04:09:33.673212 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2245b260f9b40946fb319eef185b4333c58bf0dd4e6b0b9fdcd6d0455e30ea29\": container with ID starting with 2245b260f9b40946fb319eef185b4333c58bf0dd4e6b0b9fdcd6d0455e30ea29 not found: ID does not exist" containerID="2245b260f9b40946fb319eef185b4333c58bf0dd4e6b0b9fdcd6d0455e30ea29" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.673250 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2245b260f9b40946fb319eef185b4333c58bf0dd4e6b0b9fdcd6d0455e30ea29"} err="failed to get container status \"2245b260f9b40946fb319eef185b4333c58bf0dd4e6b0b9fdcd6d0455e30ea29\": rpc error: code = NotFound desc = could not find container \"2245b260f9b40946fb319eef185b4333c58bf0dd4e6b0b9fdcd6d0455e30ea29\": container with ID starting with 2245b260f9b40946fb319eef185b4333c58bf0dd4e6b0b9fdcd6d0455e30ea29 not found: ID does not exist" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.673271 4967 scope.go:117] "RemoveContainer" containerID="e653793c0d8eee960244df6319727358785d02b58f4017b8f097cd5901bad763" Oct 11 04:09:33 crc kubenswrapper[4967]: E1011 04:09:33.673517 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e653793c0d8eee960244df6319727358785d02b58f4017b8f097cd5901bad763\": container with ID starting with e653793c0d8eee960244df6319727358785d02b58f4017b8f097cd5901bad763 not found: ID does not exist" containerID="e653793c0d8eee960244df6319727358785d02b58f4017b8f097cd5901bad763" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.673547 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e653793c0d8eee960244df6319727358785d02b58f4017b8f097cd5901bad763"} err="failed to get container status \"e653793c0d8eee960244df6319727358785d02b58f4017b8f097cd5901bad763\": rpc error: code = NotFound desc = could not find container \"e653793c0d8eee960244df6319727358785d02b58f4017b8f097cd5901bad763\": container with ID starting with e653793c0d8eee960244df6319727358785d02b58f4017b8f097cd5901bad763 not found: ID does not exist" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.724814 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.746487 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.764964 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:09:33 crc kubenswrapper[4967]: E1011 04:09:33.765445 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2121826a-fcb0-4507-ab1e-b2309a121f2d" containerName="sg-core" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.765462 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="2121826a-fcb0-4507-ab1e-b2309a121f2d" containerName="sg-core" Oct 11 04:09:33 crc kubenswrapper[4967]: E1011 04:09:33.765477 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2121826a-fcb0-4507-ab1e-b2309a121f2d" containerName="ceilometer-central-agent" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.765483 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="2121826a-fcb0-4507-ab1e-b2309a121f2d" containerName="ceilometer-central-agent" Oct 11 04:09:33 crc kubenswrapper[4967]: E1011 04:09:33.765494 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2121826a-fcb0-4507-ab1e-b2309a121f2d" containerName="ceilometer-notification-agent" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.765501 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="2121826a-fcb0-4507-ab1e-b2309a121f2d" containerName="ceilometer-notification-agent" Oct 11 04:09:33 crc kubenswrapper[4967]: E1011 04:09:33.765515 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2121826a-fcb0-4507-ab1e-b2309a121f2d" containerName="proxy-httpd" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.765522 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="2121826a-fcb0-4507-ab1e-b2309a121f2d" containerName="proxy-httpd" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.765694 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="2121826a-fcb0-4507-ab1e-b2309a121f2d" containerName="proxy-httpd" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.765714 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="2121826a-fcb0-4507-ab1e-b2309a121f2d" containerName="ceilometer-notification-agent" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.765724 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="2121826a-fcb0-4507-ab1e-b2309a121f2d" containerName="ceilometer-central-agent" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.765735 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="2121826a-fcb0-4507-ab1e-b2309a121f2d" containerName="sg-core" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.767683 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.770798 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.770824 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.771029 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.815057 4967 scope.go:117] "RemoveContainer" containerID="93bc6f25445cba6d1e4f70f9b3f209e2f24e716693d78b9c1971de07282202ed" Oct 11 04:09:33 crc kubenswrapper[4967]: E1011 04:09:33.815487 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ironic-neutron-agent\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ironic-neutron-agent pod=ironic-neutron-agent-657f6548d9-9fb8w_openstack(4390a950-7185-4fc0-b152-b5927940032b)\"" pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" podUID="4390a950-7185-4fc0-b152-b5927940032b" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.820755 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.924118 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e8932f91-bd0a-49ff-b44a-742c05a82bac-log-httpd\") pod \"ceilometer-0\" (UID: \"e8932f91-bd0a-49ff-b44a-742c05a82bac\") " pod="openstack/ceilometer-0" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.924168 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8932f91-bd0a-49ff-b44a-742c05a82bac-scripts\") pod \"ceilometer-0\" (UID: \"e8932f91-bd0a-49ff-b44a-742c05a82bac\") " pod="openstack/ceilometer-0" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.924182 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e8932f91-bd0a-49ff-b44a-742c05a82bac-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e8932f91-bd0a-49ff-b44a-742c05a82bac\") " pod="openstack/ceilometer-0" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.924202 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e8932f91-bd0a-49ff-b44a-742c05a82bac-run-httpd\") pod \"ceilometer-0\" (UID: \"e8932f91-bd0a-49ff-b44a-742c05a82bac\") " pod="openstack/ceilometer-0" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.924262 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8932f91-bd0a-49ff-b44a-742c05a82bac-config-data\") pod \"ceilometer-0\" (UID: \"e8932f91-bd0a-49ff-b44a-742c05a82bac\") " pod="openstack/ceilometer-0" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.924289 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sdnmd\" (UniqueName: \"kubernetes.io/projected/e8932f91-bd0a-49ff-b44a-742c05a82bac-kube-api-access-sdnmd\") pod \"ceilometer-0\" (UID: \"e8932f91-bd0a-49ff-b44a-742c05a82bac\") " pod="openstack/ceilometer-0" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.924331 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8932f91-bd0a-49ff-b44a-742c05a82bac-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"e8932f91-bd0a-49ff-b44a-742c05a82bac\") " pod="openstack/ceilometer-0" Oct 11 04:09:33 crc kubenswrapper[4967]: I1011 04:09:33.924359 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8932f91-bd0a-49ff-b44a-742c05a82bac-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e8932f91-bd0a-49ff-b44a-742c05a82bac\") " pod="openstack/ceilometer-0" Oct 11 04:09:34 crc kubenswrapper[4967]: I1011 04:09:34.025785 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8932f91-bd0a-49ff-b44a-742c05a82bac-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e8932f91-bd0a-49ff-b44a-742c05a82bac\") " pod="openstack/ceilometer-0" Oct 11 04:09:34 crc kubenswrapper[4967]: I1011 04:09:34.025910 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e8932f91-bd0a-49ff-b44a-742c05a82bac-log-httpd\") pod \"ceilometer-0\" (UID: \"e8932f91-bd0a-49ff-b44a-742c05a82bac\") " pod="openstack/ceilometer-0" Oct 11 04:09:34 crc kubenswrapper[4967]: I1011 04:09:34.025947 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8932f91-bd0a-49ff-b44a-742c05a82bac-scripts\") pod \"ceilometer-0\" (UID: \"e8932f91-bd0a-49ff-b44a-742c05a82bac\") " pod="openstack/ceilometer-0" Oct 11 04:09:34 crc kubenswrapper[4967]: I1011 04:09:34.025970 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e8932f91-bd0a-49ff-b44a-742c05a82bac-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e8932f91-bd0a-49ff-b44a-742c05a82bac\") " pod="openstack/ceilometer-0" Oct 11 04:09:34 crc kubenswrapper[4967]: I1011 04:09:34.026001 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e8932f91-bd0a-49ff-b44a-742c05a82bac-run-httpd\") pod \"ceilometer-0\" (UID: \"e8932f91-bd0a-49ff-b44a-742c05a82bac\") " pod="openstack/ceilometer-0" Oct 11 04:09:34 crc kubenswrapper[4967]: I1011 04:09:34.026073 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8932f91-bd0a-49ff-b44a-742c05a82bac-config-data\") pod \"ceilometer-0\" (UID: \"e8932f91-bd0a-49ff-b44a-742c05a82bac\") " pod="openstack/ceilometer-0" Oct 11 04:09:34 crc kubenswrapper[4967]: I1011 04:09:34.026134 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sdnmd\" (UniqueName: \"kubernetes.io/projected/e8932f91-bd0a-49ff-b44a-742c05a82bac-kube-api-access-sdnmd\") pod \"ceilometer-0\" (UID: \"e8932f91-bd0a-49ff-b44a-742c05a82bac\") " pod="openstack/ceilometer-0" Oct 11 04:09:34 crc kubenswrapper[4967]: I1011 04:09:34.026186 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8932f91-bd0a-49ff-b44a-742c05a82bac-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"e8932f91-bd0a-49ff-b44a-742c05a82bac\") " pod="openstack/ceilometer-0" Oct 11 04:09:34 crc kubenswrapper[4967]: I1011 04:09:34.026417 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e8932f91-bd0a-49ff-b44a-742c05a82bac-log-httpd\") pod \"ceilometer-0\" (UID: \"e8932f91-bd0a-49ff-b44a-742c05a82bac\") " pod="openstack/ceilometer-0" Oct 11 04:09:34 crc kubenswrapper[4967]: I1011 04:09:34.026692 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e8932f91-bd0a-49ff-b44a-742c05a82bac-run-httpd\") pod \"ceilometer-0\" (UID: \"e8932f91-bd0a-49ff-b44a-742c05a82bac\") " pod="openstack/ceilometer-0" Oct 11 04:09:34 crc kubenswrapper[4967]: I1011 04:09:34.030345 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e8932f91-bd0a-49ff-b44a-742c05a82bac-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"e8932f91-bd0a-49ff-b44a-742c05a82bac\") " pod="openstack/ceilometer-0" Oct 11 04:09:34 crc kubenswrapper[4967]: I1011 04:09:34.030657 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8932f91-bd0a-49ff-b44a-742c05a82bac-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"e8932f91-bd0a-49ff-b44a-742c05a82bac\") " pod="openstack/ceilometer-0" Oct 11 04:09:34 crc kubenswrapper[4967]: I1011 04:09:34.030909 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8932f91-bd0a-49ff-b44a-742c05a82bac-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"e8932f91-bd0a-49ff-b44a-742c05a82bac\") " pod="openstack/ceilometer-0" Oct 11 04:09:34 crc kubenswrapper[4967]: I1011 04:09:34.031788 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8932f91-bd0a-49ff-b44a-742c05a82bac-config-data\") pod \"ceilometer-0\" (UID: \"e8932f91-bd0a-49ff-b44a-742c05a82bac\") " pod="openstack/ceilometer-0" Oct 11 04:09:34 crc kubenswrapper[4967]: I1011 04:09:34.044485 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8932f91-bd0a-49ff-b44a-742c05a82bac-scripts\") pod \"ceilometer-0\" (UID: \"e8932f91-bd0a-49ff-b44a-742c05a82bac\") " pod="openstack/ceilometer-0" Oct 11 04:09:34 crc kubenswrapper[4967]: I1011 04:09:34.048223 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sdnmd\" (UniqueName: \"kubernetes.io/projected/e8932f91-bd0a-49ff-b44a-742c05a82bac-kube-api-access-sdnmd\") pod \"ceilometer-0\" (UID: \"e8932f91-bd0a-49ff-b44a-742c05a82bac\") " pod="openstack/ceilometer-0" Oct 11 04:09:34 crc kubenswrapper[4967]: I1011 04:09:34.091531 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:09:34 crc kubenswrapper[4967]: I1011 04:09:34.582612 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:09:34 crc kubenswrapper[4967]: W1011 04:09:34.591711 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pode8932f91_bd0a_49ff_b44a_742c05a82bac.slice/crio-fdb9907de8e086e881db9fb4ae38cf1a31f85aeb53e2db618402e75dbb8546e4 WatchSource:0}: Error finding container fdb9907de8e086e881db9fb4ae38cf1a31f85aeb53e2db618402e75dbb8546e4: Status 404 returned error can't find the container with id fdb9907de8e086e881db9fb4ae38cf1a31f85aeb53e2db618402e75dbb8546e4 Oct 11 04:09:34 crc kubenswrapper[4967]: I1011 04:09:34.824788 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2121826a-fcb0-4507-ab1e-b2309a121f2d" path="/var/lib/kubelet/pods/2121826a-fcb0-4507-ab1e-b2309a121f2d/volumes" Oct 11 04:09:35 crc kubenswrapper[4967]: I1011 04:09:35.400338 4967 generic.go:334] "Generic (PLEG): container finished" podID="d4a234ae-b210-4ac3-9bf2-841f586ee4a3" containerID="a31acefb52ee3f66f9f084ee9422d5cd5026dded84e8d4de97d3d86de40ab227" exitCode=0 Oct 11 04:09:35 crc kubenswrapper[4967]: I1011 04:09:35.400421 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xh6jw" event={"ID":"d4a234ae-b210-4ac3-9bf2-841f586ee4a3","Type":"ContainerDied","Data":"a31acefb52ee3f66f9f084ee9422d5cd5026dded84e8d4de97d3d86de40ab227"} Oct 11 04:09:35 crc kubenswrapper[4967]: I1011 04:09:35.402693 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e8932f91-bd0a-49ff-b44a-742c05a82bac","Type":"ContainerStarted","Data":"305eab7ff256536b042c2ded7316bc8c9d23123505674310af3a13e2fee28a04"} Oct 11 04:09:35 crc kubenswrapper[4967]: I1011 04:09:35.402733 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e8932f91-bd0a-49ff-b44a-742c05a82bac","Type":"ContainerStarted","Data":"fdb9907de8e086e881db9fb4ae38cf1a31f85aeb53e2db618402e75dbb8546e4"} Oct 11 04:09:36 crc kubenswrapper[4967]: I1011 04:09:36.414422 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e8932f91-bd0a-49ff-b44a-742c05a82bac","Type":"ContainerStarted","Data":"f288f48b8ef8aa9f696a381c0911c3979d4348a047e4d1efb12615ab4900d656"} Oct 11 04:09:36 crc kubenswrapper[4967]: I1011 04:09:36.788121 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xh6jw" Oct 11 04:09:36 crc kubenswrapper[4967]: I1011 04:09:36.877030 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bxp55\" (UniqueName: \"kubernetes.io/projected/d4a234ae-b210-4ac3-9bf2-841f586ee4a3-kube-api-access-bxp55\") pod \"d4a234ae-b210-4ac3-9bf2-841f586ee4a3\" (UID: \"d4a234ae-b210-4ac3-9bf2-841f586ee4a3\") " Oct 11 04:09:36 crc kubenswrapper[4967]: I1011 04:09:36.877101 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4a234ae-b210-4ac3-9bf2-841f586ee4a3-scripts\") pod \"d4a234ae-b210-4ac3-9bf2-841f586ee4a3\" (UID: \"d4a234ae-b210-4ac3-9bf2-841f586ee4a3\") " Oct 11 04:09:36 crc kubenswrapper[4967]: I1011 04:09:36.877339 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4a234ae-b210-4ac3-9bf2-841f586ee4a3-combined-ca-bundle\") pod \"d4a234ae-b210-4ac3-9bf2-841f586ee4a3\" (UID: \"d4a234ae-b210-4ac3-9bf2-841f586ee4a3\") " Oct 11 04:09:36 crc kubenswrapper[4967]: I1011 04:09:36.877372 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4a234ae-b210-4ac3-9bf2-841f586ee4a3-config-data\") pod \"d4a234ae-b210-4ac3-9bf2-841f586ee4a3\" (UID: \"d4a234ae-b210-4ac3-9bf2-841f586ee4a3\") " Oct 11 04:09:36 crc kubenswrapper[4967]: I1011 04:09:36.882641 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4a234ae-b210-4ac3-9bf2-841f586ee4a3-scripts" (OuterVolumeSpecName: "scripts") pod "d4a234ae-b210-4ac3-9bf2-841f586ee4a3" (UID: "d4a234ae-b210-4ac3-9bf2-841f586ee4a3"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:09:36 crc kubenswrapper[4967]: I1011 04:09:36.883886 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d4a234ae-b210-4ac3-9bf2-841f586ee4a3-kube-api-access-bxp55" (OuterVolumeSpecName: "kube-api-access-bxp55") pod "d4a234ae-b210-4ac3-9bf2-841f586ee4a3" (UID: "d4a234ae-b210-4ac3-9bf2-841f586ee4a3"). InnerVolumeSpecName "kube-api-access-bxp55". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:09:36 crc kubenswrapper[4967]: I1011 04:09:36.914585 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4a234ae-b210-4ac3-9bf2-841f586ee4a3-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "d4a234ae-b210-4ac3-9bf2-841f586ee4a3" (UID: "d4a234ae-b210-4ac3-9bf2-841f586ee4a3"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:09:36 crc kubenswrapper[4967]: I1011 04:09:36.926289 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/d4a234ae-b210-4ac3-9bf2-841f586ee4a3-config-data" (OuterVolumeSpecName: "config-data") pod "d4a234ae-b210-4ac3-9bf2-841f586ee4a3" (UID: "d4a234ae-b210-4ac3-9bf2-841f586ee4a3"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:09:36 crc kubenswrapper[4967]: I1011 04:09:36.980000 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bxp55\" (UniqueName: \"kubernetes.io/projected/d4a234ae-b210-4ac3-9bf2-841f586ee4a3-kube-api-access-bxp55\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:36 crc kubenswrapper[4967]: I1011 04:09:36.980047 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/d4a234ae-b210-4ac3-9bf2-841f586ee4a3-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:36 crc kubenswrapper[4967]: I1011 04:09:36.980057 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/d4a234ae-b210-4ac3-9bf2-841f586ee4a3-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:36 crc kubenswrapper[4967]: I1011 04:09:36.980065 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/d4a234ae-b210-4ac3-9bf2-841f586ee4a3-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:37 crc kubenswrapper[4967]: I1011 04:09:37.425279 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-db-sync-xh6jw" event={"ID":"d4a234ae-b210-4ac3-9bf2-841f586ee4a3","Type":"ContainerDied","Data":"5a2c3abcc638399339b562559ac6368d2e3d72c071d2391fc2242f4807c68e60"} Oct 11 04:09:37 crc kubenswrapper[4967]: I1011 04:09:37.425320 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5a2c3abcc638399339b562559ac6368d2e3d72c071d2391fc2242f4807c68e60" Oct 11 04:09:37 crc kubenswrapper[4967]: I1011 04:09:37.425334 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-db-sync-xh6jw" Oct 11 04:09:37 crc kubenswrapper[4967]: I1011 04:09:37.625159 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 11 04:09:37 crc kubenswrapper[4967]: E1011 04:09:37.625563 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d4a234ae-b210-4ac3-9bf2-841f586ee4a3" containerName="nova-cell0-conductor-db-sync" Oct 11 04:09:37 crc kubenswrapper[4967]: I1011 04:09:37.625580 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="d4a234ae-b210-4ac3-9bf2-841f586ee4a3" containerName="nova-cell0-conductor-db-sync" Oct 11 04:09:37 crc kubenswrapper[4967]: I1011 04:09:37.625872 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="d4a234ae-b210-4ac3-9bf2-841f586ee4a3" containerName="nova-cell0-conductor-db-sync" Oct 11 04:09:37 crc kubenswrapper[4967]: I1011 04:09:37.626483 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 11 04:09:37 crc kubenswrapper[4967]: I1011 04:09:37.634587 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-nova-dockercfg-fj2pf" Oct 11 04:09:37 crc kubenswrapper[4967]: I1011 04:09:37.635122 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-conductor-config-data" Oct 11 04:09:37 crc kubenswrapper[4967]: I1011 04:09:37.635669 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 11 04:09:37 crc kubenswrapper[4967]: I1011 04:09:37.797701 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67b30c76-971c-4f10-a692-4bbf2a1e8531-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"67b30c76-971c-4f10-a692-4bbf2a1e8531\") " pod="openstack/nova-cell0-conductor-0" Oct 11 04:09:37 crc kubenswrapper[4967]: I1011 04:09:37.797882 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67b30c76-971c-4f10-a692-4bbf2a1e8531-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"67b30c76-971c-4f10-a692-4bbf2a1e8531\") " pod="openstack/nova-cell0-conductor-0" Oct 11 04:09:37 crc kubenswrapper[4967]: I1011 04:09:37.797921 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wr4jp\" (UniqueName: \"kubernetes.io/projected/67b30c76-971c-4f10-a692-4bbf2a1e8531-kube-api-access-wr4jp\") pod \"nova-cell0-conductor-0\" (UID: \"67b30c76-971c-4f10-a692-4bbf2a1e8531\") " pod="openstack/nova-cell0-conductor-0" Oct 11 04:09:37 crc kubenswrapper[4967]: I1011 04:09:37.899883 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67b30c76-971c-4f10-a692-4bbf2a1e8531-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"67b30c76-971c-4f10-a692-4bbf2a1e8531\") " pod="openstack/nova-cell0-conductor-0" Oct 11 04:09:37 crc kubenswrapper[4967]: I1011 04:09:37.900053 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67b30c76-971c-4f10-a692-4bbf2a1e8531-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"67b30c76-971c-4f10-a692-4bbf2a1e8531\") " pod="openstack/nova-cell0-conductor-0" Oct 11 04:09:37 crc kubenswrapper[4967]: I1011 04:09:37.900167 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wr4jp\" (UniqueName: \"kubernetes.io/projected/67b30c76-971c-4f10-a692-4bbf2a1e8531-kube-api-access-wr4jp\") pod \"nova-cell0-conductor-0\" (UID: \"67b30c76-971c-4f10-a692-4bbf2a1e8531\") " pod="openstack/nova-cell0-conductor-0" Oct 11 04:09:37 crc kubenswrapper[4967]: I1011 04:09:37.908553 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/67b30c76-971c-4f10-a692-4bbf2a1e8531-config-data\") pod \"nova-cell0-conductor-0\" (UID: \"67b30c76-971c-4f10-a692-4bbf2a1e8531\") " pod="openstack/nova-cell0-conductor-0" Oct 11 04:09:37 crc kubenswrapper[4967]: I1011 04:09:37.919904 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/67b30c76-971c-4f10-a692-4bbf2a1e8531-combined-ca-bundle\") pod \"nova-cell0-conductor-0\" (UID: \"67b30c76-971c-4f10-a692-4bbf2a1e8531\") " pod="openstack/nova-cell0-conductor-0" Oct 11 04:09:37 crc kubenswrapper[4967]: I1011 04:09:37.921743 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wr4jp\" (UniqueName: \"kubernetes.io/projected/67b30c76-971c-4f10-a692-4bbf2a1e8531-kube-api-access-wr4jp\") pod \"nova-cell0-conductor-0\" (UID: \"67b30c76-971c-4f10-a692-4bbf2a1e8531\") " pod="openstack/nova-cell0-conductor-0" Oct 11 04:09:37 crc kubenswrapper[4967]: I1011 04:09:37.974802 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-conductor-0" Oct 11 04:09:38 crc kubenswrapper[4967]: I1011 04:09:38.469013 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-conductor-0"] Oct 11 04:09:38 crc kubenswrapper[4967]: W1011 04:09:38.471763 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod67b30c76_971c_4f10_a692_4bbf2a1e8531.slice/crio-232d59f703524fb97a5c9d102386a64feaf7951ba4c33d16907a563b121b790f WatchSource:0}: Error finding container 232d59f703524fb97a5c9d102386a64feaf7951ba4c33d16907a563b121b790f: Status 404 returned error can't find the container with id 232d59f703524fb97a5c9d102386a64feaf7951ba4c33d16907a563b121b790f Oct 11 04:09:39 crc kubenswrapper[4967]: I1011 04:09:39.450543 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e8932f91-bd0a-49ff-b44a-742c05a82bac","Type":"ContainerStarted","Data":"c627151d2b9fff4f40ac8e23f089cc1723101cfa1e278c02fa45ca4ddd5a8c5d"} Oct 11 04:09:39 crc kubenswrapper[4967]: I1011 04:09:39.453305 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"67b30c76-971c-4f10-a692-4bbf2a1e8531","Type":"ContainerStarted","Data":"232d59f703524fb97a5c9d102386a64feaf7951ba4c33d16907a563b121b790f"} Oct 11 04:09:40 crc kubenswrapper[4967]: I1011 04:09:40.467237 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e8932f91-bd0a-49ff-b44a-742c05a82bac","Type":"ContainerStarted","Data":"e5199f2b9b10c94ccb043c04b087750e60d26ad8f43f3e750553ea6d5c3bddad"} Oct 11 04:09:40 crc kubenswrapper[4967]: I1011 04:09:40.468094 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 11 04:09:40 crc kubenswrapper[4967]: I1011 04:09:40.471800 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-conductor-0" event={"ID":"67b30c76-971c-4f10-a692-4bbf2a1e8531","Type":"ContainerStarted","Data":"fe10a95f69bdcab6b457cceb97f43712a26a80ab450a848f65154d1639ce2cd4"} Oct 11 04:09:40 crc kubenswrapper[4967]: I1011 04:09:40.471932 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell0-conductor-0" Oct 11 04:09:40 crc kubenswrapper[4967]: I1011 04:09:40.500981 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=1.983640098 podStartE2EDuration="7.500964056s" podCreationTimestamp="2025-10-11 04:09:33 +0000 UTC" firstStartedPulling="2025-10-11 04:09:34.595383576 +0000 UTC m=+1102.558592509" lastFinishedPulling="2025-10-11 04:09:40.112707534 +0000 UTC m=+1108.075916467" observedRunningTime="2025-10-11 04:09:40.496068002 +0000 UTC m=+1108.459276935" watchObservedRunningTime="2025-10-11 04:09:40.500964056 +0000 UTC m=+1108.464172989" Oct 11 04:09:40 crc kubenswrapper[4967]: I1011 04:09:40.525016 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-conductor-0" podStartSLOduration=3.524992322 podStartE2EDuration="3.524992322s" podCreationTimestamp="2025-10-11 04:09:37 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:09:40.518278816 +0000 UTC m=+1108.481487749" watchObservedRunningTime="2025-10-11 04:09:40.524992322 +0000 UTC m=+1108.488201255" Oct 11 04:09:42 crc kubenswrapper[4967]: I1011 04:09:42.993438 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:09:42 crc kubenswrapper[4967]: I1011 04:09:42.994193 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e8932f91-bd0a-49ff-b44a-742c05a82bac" containerName="ceilometer-central-agent" containerID="cri-o://305eab7ff256536b042c2ded7316bc8c9d23123505674310af3a13e2fee28a04" gracePeriod=30 Oct 11 04:09:42 crc kubenswrapper[4967]: I1011 04:09:42.994646 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e8932f91-bd0a-49ff-b44a-742c05a82bac" containerName="proxy-httpd" containerID="cri-o://e5199f2b9b10c94ccb043c04b087750e60d26ad8f43f3e750553ea6d5c3bddad" gracePeriod=30 Oct 11 04:09:42 crc kubenswrapper[4967]: I1011 04:09:42.994705 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e8932f91-bd0a-49ff-b44a-742c05a82bac" containerName="sg-core" containerID="cri-o://c627151d2b9fff4f40ac8e23f089cc1723101cfa1e278c02fa45ca4ddd5a8c5d" gracePeriod=30 Oct 11 04:09:42 crc kubenswrapper[4967]: I1011 04:09:42.994747 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="e8932f91-bd0a-49ff-b44a-742c05a82bac" containerName="ceilometer-notification-agent" containerID="cri-o://f288f48b8ef8aa9f696a381c0911c3979d4348a047e4d1efb12615ab4900d656" gracePeriod=30 Oct 11 04:09:43 crc kubenswrapper[4967]: I1011 04:09:43.529595 4967 generic.go:334] "Generic (PLEG): container finished" podID="e8932f91-bd0a-49ff-b44a-742c05a82bac" containerID="e5199f2b9b10c94ccb043c04b087750e60d26ad8f43f3e750553ea6d5c3bddad" exitCode=0 Oct 11 04:09:43 crc kubenswrapper[4967]: I1011 04:09:43.530300 4967 generic.go:334] "Generic (PLEG): container finished" podID="e8932f91-bd0a-49ff-b44a-742c05a82bac" containerID="c627151d2b9fff4f40ac8e23f089cc1723101cfa1e278c02fa45ca4ddd5a8c5d" exitCode=2 Oct 11 04:09:43 crc kubenswrapper[4967]: I1011 04:09:43.530314 4967 generic.go:334] "Generic (PLEG): container finished" podID="e8932f91-bd0a-49ff-b44a-742c05a82bac" containerID="f288f48b8ef8aa9f696a381c0911c3979d4348a047e4d1efb12615ab4900d656" exitCode=0 Oct 11 04:09:43 crc kubenswrapper[4967]: I1011 04:09:43.530392 4967 generic.go:334] "Generic (PLEG): container finished" podID="e8932f91-bd0a-49ff-b44a-742c05a82bac" containerID="305eab7ff256536b042c2ded7316bc8c9d23123505674310af3a13e2fee28a04" exitCode=0 Oct 11 04:09:43 crc kubenswrapper[4967]: I1011 04:09:43.529772 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e8932f91-bd0a-49ff-b44a-742c05a82bac","Type":"ContainerDied","Data":"e5199f2b9b10c94ccb043c04b087750e60d26ad8f43f3e750553ea6d5c3bddad"} Oct 11 04:09:43 crc kubenswrapper[4967]: I1011 04:09:43.530478 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e8932f91-bd0a-49ff-b44a-742c05a82bac","Type":"ContainerDied","Data":"c627151d2b9fff4f40ac8e23f089cc1723101cfa1e278c02fa45ca4ddd5a8c5d"} Oct 11 04:09:43 crc kubenswrapper[4967]: I1011 04:09:43.530493 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e8932f91-bd0a-49ff-b44a-742c05a82bac","Type":"ContainerDied","Data":"f288f48b8ef8aa9f696a381c0911c3979d4348a047e4d1efb12615ab4900d656"} Oct 11 04:09:43 crc kubenswrapper[4967]: I1011 04:09:43.530504 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e8932f91-bd0a-49ff-b44a-742c05a82bac","Type":"ContainerDied","Data":"305eab7ff256536b042c2ded7316bc8c9d23123505674310af3a13e2fee28a04"} Oct 11 04:09:43 crc kubenswrapper[4967]: I1011 04:09:43.733883 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:09:43 crc kubenswrapper[4967]: I1011 04:09:43.849049 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8932f91-bd0a-49ff-b44a-742c05a82bac-ceilometer-tls-certs\") pod \"e8932f91-bd0a-49ff-b44a-742c05a82bac\" (UID: \"e8932f91-bd0a-49ff-b44a-742c05a82bac\") " Oct 11 04:09:43 crc kubenswrapper[4967]: I1011 04:09:43.849887 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8932f91-bd0a-49ff-b44a-742c05a82bac-combined-ca-bundle\") pod \"e8932f91-bd0a-49ff-b44a-742c05a82bac\" (UID: \"e8932f91-bd0a-49ff-b44a-742c05a82bac\") " Oct 11 04:09:43 crc kubenswrapper[4967]: I1011 04:09:43.849940 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e8932f91-bd0a-49ff-b44a-742c05a82bac-run-httpd\") pod \"e8932f91-bd0a-49ff-b44a-742c05a82bac\" (UID: \"e8932f91-bd0a-49ff-b44a-742c05a82bac\") " Oct 11 04:09:43 crc kubenswrapper[4967]: I1011 04:09:43.849969 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8932f91-bd0a-49ff-b44a-742c05a82bac-config-data\") pod \"e8932f91-bd0a-49ff-b44a-742c05a82bac\" (UID: \"e8932f91-bd0a-49ff-b44a-742c05a82bac\") " Oct 11 04:09:43 crc kubenswrapper[4967]: I1011 04:09:43.850022 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sdnmd\" (UniqueName: \"kubernetes.io/projected/e8932f91-bd0a-49ff-b44a-742c05a82bac-kube-api-access-sdnmd\") pod \"e8932f91-bd0a-49ff-b44a-742c05a82bac\" (UID: \"e8932f91-bd0a-49ff-b44a-742c05a82bac\") " Oct 11 04:09:43 crc kubenswrapper[4967]: I1011 04:09:43.850050 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e8932f91-bd0a-49ff-b44a-742c05a82bac-sg-core-conf-yaml\") pod \"e8932f91-bd0a-49ff-b44a-742c05a82bac\" (UID: \"e8932f91-bd0a-49ff-b44a-742c05a82bac\") " Oct 11 04:09:43 crc kubenswrapper[4967]: I1011 04:09:43.850220 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e8932f91-bd0a-49ff-b44a-742c05a82bac-log-httpd\") pod \"e8932f91-bd0a-49ff-b44a-742c05a82bac\" (UID: \"e8932f91-bd0a-49ff-b44a-742c05a82bac\") " Oct 11 04:09:43 crc kubenswrapper[4967]: I1011 04:09:43.850272 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8932f91-bd0a-49ff-b44a-742c05a82bac-scripts\") pod \"e8932f91-bd0a-49ff-b44a-742c05a82bac\" (UID: \"e8932f91-bd0a-49ff-b44a-742c05a82bac\") " Oct 11 04:09:43 crc kubenswrapper[4967]: I1011 04:09:43.850362 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8932f91-bd0a-49ff-b44a-742c05a82bac-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "e8932f91-bd0a-49ff-b44a-742c05a82bac" (UID: "e8932f91-bd0a-49ff-b44a-742c05a82bac"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:09:43 crc kubenswrapper[4967]: I1011 04:09:43.850609 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e8932f91-bd0a-49ff-b44a-742c05a82bac-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "e8932f91-bd0a-49ff-b44a-742c05a82bac" (UID: "e8932f91-bd0a-49ff-b44a-742c05a82bac"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:09:43 crc kubenswrapper[4967]: I1011 04:09:43.850948 4967 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e8932f91-bd0a-49ff-b44a-742c05a82bac-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:43 crc kubenswrapper[4967]: I1011 04:09:43.850966 4967 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/e8932f91-bd0a-49ff-b44a-742c05a82bac-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:43 crc kubenswrapper[4967]: I1011 04:09:43.854393 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e8932f91-bd0a-49ff-b44a-742c05a82bac-kube-api-access-sdnmd" (OuterVolumeSpecName: "kube-api-access-sdnmd") pod "e8932f91-bd0a-49ff-b44a-742c05a82bac" (UID: "e8932f91-bd0a-49ff-b44a-742c05a82bac"). InnerVolumeSpecName "kube-api-access-sdnmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:09:43 crc kubenswrapper[4967]: I1011 04:09:43.855729 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8932f91-bd0a-49ff-b44a-742c05a82bac-scripts" (OuterVolumeSpecName: "scripts") pod "e8932f91-bd0a-49ff-b44a-742c05a82bac" (UID: "e8932f91-bd0a-49ff-b44a-742c05a82bac"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:09:43 crc kubenswrapper[4967]: I1011 04:09:43.880033 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8932f91-bd0a-49ff-b44a-742c05a82bac-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "e8932f91-bd0a-49ff-b44a-742c05a82bac" (UID: "e8932f91-bd0a-49ff-b44a-742c05a82bac"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:09:43 crc kubenswrapper[4967]: I1011 04:09:43.903236 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8932f91-bd0a-49ff-b44a-742c05a82bac-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "e8932f91-bd0a-49ff-b44a-742c05a82bac" (UID: "e8932f91-bd0a-49ff-b44a-742c05a82bac"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:09:43 crc kubenswrapper[4967]: I1011 04:09:43.925611 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8932f91-bd0a-49ff-b44a-742c05a82bac-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "e8932f91-bd0a-49ff-b44a-742c05a82bac" (UID: "e8932f91-bd0a-49ff-b44a-742c05a82bac"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:09:43 crc kubenswrapper[4967]: I1011 04:09:43.953168 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/e8932f91-bd0a-49ff-b44a-742c05a82bac-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:43 crc kubenswrapper[4967]: I1011 04:09:43.953202 4967 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/e8932f91-bd0a-49ff-b44a-742c05a82bac-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:43 crc kubenswrapper[4967]: I1011 04:09:43.953212 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/e8932f91-bd0a-49ff-b44a-742c05a82bac-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:43 crc kubenswrapper[4967]: I1011 04:09:43.953221 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sdnmd\" (UniqueName: \"kubernetes.io/projected/e8932f91-bd0a-49ff-b44a-742c05a82bac-kube-api-access-sdnmd\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:43 crc kubenswrapper[4967]: I1011 04:09:43.953230 4967 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/e8932f91-bd0a-49ff-b44a-742c05a82bac-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:43 crc kubenswrapper[4967]: I1011 04:09:43.958228 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e8932f91-bd0a-49ff-b44a-742c05a82bac-config-data" (OuterVolumeSpecName: "config-data") pod "e8932f91-bd0a-49ff-b44a-742c05a82bac" (UID: "e8932f91-bd0a-49ff-b44a-742c05a82bac"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.055613 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/e8932f91-bd0a-49ff-b44a-742c05a82bac-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.089418 4967 patch_prober.go:28] interesting pod/machine-config-daemon-sgrv2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.089486 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.543435 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"e8932f91-bd0a-49ff-b44a-742c05a82bac","Type":"ContainerDied","Data":"fdb9907de8e086e881db9fb4ae38cf1a31f85aeb53e2db618402e75dbb8546e4"} Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.543532 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.543542 4967 scope.go:117] "RemoveContainer" containerID="e5199f2b9b10c94ccb043c04b087750e60d26ad8f43f3e750553ea6d5c3bddad" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.573402 4967 scope.go:117] "RemoveContainer" containerID="c627151d2b9fff4f40ac8e23f089cc1723101cfa1e278c02fa45ca4ddd5a8c5d" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.597006 4967 scope.go:117] "RemoveContainer" containerID="f288f48b8ef8aa9f696a381c0911c3979d4348a047e4d1efb12615ab4900d656" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.600210 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.614508 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.626102 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:09:44 crc kubenswrapper[4967]: E1011 04:09:44.626855 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8932f91-bd0a-49ff-b44a-742c05a82bac" containerName="ceilometer-central-agent" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.626884 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8932f91-bd0a-49ff-b44a-742c05a82bac" containerName="ceilometer-central-agent" Oct 11 04:09:44 crc kubenswrapper[4967]: E1011 04:09:44.626911 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8932f91-bd0a-49ff-b44a-742c05a82bac" containerName="ceilometer-notification-agent" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.626925 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8932f91-bd0a-49ff-b44a-742c05a82bac" containerName="ceilometer-notification-agent" Oct 11 04:09:44 crc kubenswrapper[4967]: E1011 04:09:44.626964 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8932f91-bd0a-49ff-b44a-742c05a82bac" containerName="proxy-httpd" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.626979 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8932f91-bd0a-49ff-b44a-742c05a82bac" containerName="proxy-httpd" Oct 11 04:09:44 crc kubenswrapper[4967]: E1011 04:09:44.627023 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e8932f91-bd0a-49ff-b44a-742c05a82bac" containerName="sg-core" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.627036 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="e8932f91-bd0a-49ff-b44a-742c05a82bac" containerName="sg-core" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.627377 4967 scope.go:117] "RemoveContainer" containerID="305eab7ff256536b042c2ded7316bc8c9d23123505674310af3a13e2fee28a04" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.627466 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8932f91-bd0a-49ff-b44a-742c05a82bac" containerName="ceilometer-central-agent" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.627524 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8932f91-bd0a-49ff-b44a-742c05a82bac" containerName="ceilometer-notification-agent" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.627556 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8932f91-bd0a-49ff-b44a-742c05a82bac" containerName="proxy-httpd" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.627578 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="e8932f91-bd0a-49ff-b44a-742c05a82bac" containerName="sg-core" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.635545 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.635824 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.641509 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.646395 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.646940 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.769131 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-config-data\") pod \"ceilometer-0\" (UID: \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\") " pod="openstack/ceilometer-0" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.769177 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-log-httpd\") pod \"ceilometer-0\" (UID: \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\") " pod="openstack/ceilometer-0" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.769212 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\") " pod="openstack/ceilometer-0" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.769303 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\") " pod="openstack/ceilometer-0" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.769336 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-scripts\") pod \"ceilometer-0\" (UID: \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\") " pod="openstack/ceilometer-0" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.769374 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\") " pod="openstack/ceilometer-0" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.769411 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-run-httpd\") pod \"ceilometer-0\" (UID: \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\") " pod="openstack/ceilometer-0" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.769433 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-plj5z\" (UniqueName: \"kubernetes.io/projected/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-kube-api-access-plj5z\") pod \"ceilometer-0\" (UID: \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\") " pod="openstack/ceilometer-0" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.824719 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e8932f91-bd0a-49ff-b44a-742c05a82bac" path="/var/lib/kubelet/pods/e8932f91-bd0a-49ff-b44a-742c05a82bac/volumes" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.871282 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\") " pod="openstack/ceilometer-0" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.871362 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-scripts\") pod \"ceilometer-0\" (UID: \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\") " pod="openstack/ceilometer-0" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.871399 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\") " pod="openstack/ceilometer-0" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.871451 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-run-httpd\") pod \"ceilometer-0\" (UID: \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\") " pod="openstack/ceilometer-0" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.871503 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-plj5z\" (UniqueName: \"kubernetes.io/projected/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-kube-api-access-plj5z\") pod \"ceilometer-0\" (UID: \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\") " pod="openstack/ceilometer-0" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.871560 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-config-data\") pod \"ceilometer-0\" (UID: \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\") " pod="openstack/ceilometer-0" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.871590 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-log-httpd\") pod \"ceilometer-0\" (UID: \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\") " pod="openstack/ceilometer-0" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.871625 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\") " pod="openstack/ceilometer-0" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.872043 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-log-httpd\") pod \"ceilometer-0\" (UID: \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\") " pod="openstack/ceilometer-0" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.872130 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-run-httpd\") pod \"ceilometer-0\" (UID: \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\") " pod="openstack/ceilometer-0" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.877498 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-scripts\") pod \"ceilometer-0\" (UID: \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\") " pod="openstack/ceilometer-0" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.880853 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\") " pod="openstack/ceilometer-0" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.881293 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\") " pod="openstack/ceilometer-0" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.885801 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\") " pod="openstack/ceilometer-0" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.887782 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-plj5z\" (UniqueName: \"kubernetes.io/projected/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-kube-api-access-plj5z\") pod \"ceilometer-0\" (UID: \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\") " pod="openstack/ceilometer-0" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.899875 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-config-data\") pod \"ceilometer-0\" (UID: \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\") " pod="openstack/ceilometer-0" Oct 11 04:09:44 crc kubenswrapper[4967]: I1011 04:09:44.964411 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:09:45 crc kubenswrapper[4967]: I1011 04:09:45.467281 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:09:45 crc kubenswrapper[4967]: W1011 04:09:45.480625 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e95b396_4e67_4df5_a39a_15ff3a00e6a1.slice/crio-e226bc76f956e91007e49bc6d6a7b72a16216c26fa9caa5ec70ba33caa30e164 WatchSource:0}: Error finding container e226bc76f956e91007e49bc6d6a7b72a16216c26fa9caa5ec70ba33caa30e164: Status 404 returned error can't find the container with id e226bc76f956e91007e49bc6d6a7b72a16216c26fa9caa5ec70ba33caa30e164 Oct 11 04:09:45 crc kubenswrapper[4967]: I1011 04:09:45.558366 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e95b396-4e67-4df5-a39a-15ff3a00e6a1","Type":"ContainerStarted","Data":"e226bc76f956e91007e49bc6d6a7b72a16216c26fa9caa5ec70ba33caa30e164"} Oct 11 04:09:46 crc kubenswrapper[4967]: I1011 04:09:46.569535 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e95b396-4e67-4df5-a39a-15ff3a00e6a1","Type":"ContainerStarted","Data":"f7362d856d4a38869ca9e64c37955f1a0544ca77568782afb13d718ba74b351a"} Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.004858 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell0-conductor-0" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.470050 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell0-cell-mapping-krtvp"] Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.472150 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-krtvp" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.474547 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-config-data" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.474960 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell0-manage-scripts" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.484541 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-krtvp"] Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.543001 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e-scripts\") pod \"nova-cell0-cell-mapping-krtvp\" (UID: \"9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e\") " pod="openstack/nova-cell0-cell-mapping-krtvp" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.543058 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-krtvp\" (UID: \"9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e\") " pod="openstack/nova-cell0-cell-mapping-krtvp" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.543198 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e-config-data\") pod \"nova-cell0-cell-mapping-krtvp\" (UID: \"9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e\") " pod="openstack/nova-cell0-cell-mapping-krtvp" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.543290 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-scd22\" (UniqueName: \"kubernetes.io/projected/9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e-kube-api-access-scd22\") pod \"nova-cell0-cell-mapping-krtvp\" (UID: \"9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e\") " pod="openstack/nova-cell0-cell-mapping-krtvp" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.604776 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e95b396-4e67-4df5-a39a-15ff3a00e6a1","Type":"ContainerStarted","Data":"0875cf7cbc65df1e3e6e99f37987fe6f36adf3677c200154187ee9d1072fac73"} Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.652521 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e-scripts\") pod \"nova-cell0-cell-mapping-krtvp\" (UID: \"9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e\") " pod="openstack/nova-cell0-cell-mapping-krtvp" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.652838 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-krtvp\" (UID: \"9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e\") " pod="openstack/nova-cell0-cell-mapping-krtvp" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.652936 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e-config-data\") pod \"nova-cell0-cell-mapping-krtvp\" (UID: \"9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e\") " pod="openstack/nova-cell0-cell-mapping-krtvp" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.653030 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-scd22\" (UniqueName: \"kubernetes.io/projected/9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e-kube-api-access-scd22\") pod \"nova-cell0-cell-mapping-krtvp\" (UID: \"9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e\") " pod="openstack/nova-cell0-cell-mapping-krtvp" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.657262 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.658707 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.664891 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.672507 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e-scripts\") pod \"nova-cell0-cell-mapping-krtvp\" (UID: \"9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e\") " pod="openstack/nova-cell0-cell-mapping-krtvp" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.672726 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e-combined-ca-bundle\") pod \"nova-cell0-cell-mapping-krtvp\" (UID: \"9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e\") " pod="openstack/nova-cell0-cell-mapping-krtvp" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.678183 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.680260 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.684107 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.688312 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-scd22\" (UniqueName: \"kubernetes.io/projected/9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e-kube-api-access-scd22\") pod \"nova-cell0-cell-mapping-krtvp\" (UID: \"9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e\") " pod="openstack/nova-cell0-cell-mapping-krtvp" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.695142 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.698852 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e-config-data\") pod \"nova-cell0-cell-mapping-krtvp\" (UID: \"9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e\") " pod="openstack/nova-cell0-cell-mapping-krtvp" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.721188 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.767675 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cee26db2-5786-43ed-8a26-3b0e55ca2964-config-data\") pod \"nova-api-0\" (UID: \"cee26db2-5786-43ed-8a26-3b0e55ca2964\") " pod="openstack/nova-api-0" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.768234 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cee26db2-5786-43ed-8a26-3b0e55ca2964-logs\") pod \"nova-api-0\" (UID: \"cee26db2-5786-43ed-8a26-3b0e55ca2964\") " pod="openstack/nova-api-0" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.768426 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cee26db2-5786-43ed-8a26-3b0e55ca2964-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cee26db2-5786-43ed-8a26-3b0e55ca2964\") " pod="openstack/nova-api-0" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.768540 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-m98f2\" (UniqueName: \"kubernetes.io/projected/cee26db2-5786-43ed-8a26-3b0e55ca2964-kube-api-access-m98f2\") pod \"nova-api-0\" (UID: \"cee26db2-5786-43ed-8a26-3b0e55ca2964\") " pod="openstack/nova-api-0" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.818743 4967 scope.go:117] "RemoveContainer" containerID="93bc6f25445cba6d1e4f70f9b3f209e2f24e716693d78b9c1971de07282202ed" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.850679 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-krtvp" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.905357 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r4qbf\" (UniqueName: \"kubernetes.io/projected/1317501c-8db9-4f03-a1cf-10e92b2ccc8b-kube-api-access-r4qbf\") pod \"nova-cell1-novncproxy-0\" (UID: \"1317501c-8db9-4f03-a1cf-10e92b2ccc8b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.906162 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cee26db2-5786-43ed-8a26-3b0e55ca2964-config-data\") pod \"nova-api-0\" (UID: \"cee26db2-5786-43ed-8a26-3b0e55ca2964\") " pod="openstack/nova-api-0" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.906274 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1317501c-8db9-4f03-a1cf-10e92b2ccc8b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1317501c-8db9-4f03-a1cf-10e92b2ccc8b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.906361 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1317501c-8db9-4f03-a1cf-10e92b2ccc8b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1317501c-8db9-4f03-a1cf-10e92b2ccc8b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.906433 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cee26db2-5786-43ed-8a26-3b0e55ca2964-logs\") pod \"nova-api-0\" (UID: \"cee26db2-5786-43ed-8a26-3b0e55ca2964\") " pod="openstack/nova-api-0" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.906543 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cee26db2-5786-43ed-8a26-3b0e55ca2964-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cee26db2-5786-43ed-8a26-3b0e55ca2964\") " pod="openstack/nova-api-0" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.906580 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-m98f2\" (UniqueName: \"kubernetes.io/projected/cee26db2-5786-43ed-8a26-3b0e55ca2964-kube-api-access-m98f2\") pod \"nova-api-0\" (UID: \"cee26db2-5786-43ed-8a26-3b0e55ca2964\") " pod="openstack/nova-api-0" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.908895 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cee26db2-5786-43ed-8a26-3b0e55ca2964-logs\") pod \"nova-api-0\" (UID: \"cee26db2-5786-43ed-8a26-3b0e55ca2964\") " pod="openstack/nova-api-0" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.918720 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cee26db2-5786-43ed-8a26-3b0e55ca2964-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"cee26db2-5786-43ed-8a26-3b0e55ca2964\") " pod="openstack/nova-api-0" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.924523 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.926759 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.934230 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.942532 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cee26db2-5786-43ed-8a26-3b0e55ca2964-config-data\") pod \"nova-api-0\" (UID: \"cee26db2-5786-43ed-8a26-3b0e55ca2964\") " pod="openstack/nova-api-0" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.963924 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-m98f2\" (UniqueName: \"kubernetes.io/projected/cee26db2-5786-43ed-8a26-3b0e55ca2964-kube-api-access-m98f2\") pod \"nova-api-0\" (UID: \"cee26db2-5786-43ed-8a26-3b0e55ca2964\") " pod="openstack/nova-api-0" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.975212 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.976738 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.978965 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 11 04:09:48 crc kubenswrapper[4967]: I1011 04:09:48.996041 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.004934 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.007192 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/470abdb5-fefb-48c8-bb4c-9ad0e81caedb-logs\") pod \"nova-metadata-0\" (UID: \"470abdb5-fefb-48c8-bb4c-9ad0e81caedb\") " pod="openstack/nova-metadata-0" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.008197 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ggr79\" (UniqueName: \"kubernetes.io/projected/7fd48f8c-e380-44cb-8c5b-8380b637e0ef-kube-api-access-ggr79\") pod \"nova-scheduler-0\" (UID: \"7fd48f8c-e380-44cb-8c5b-8380b637e0ef\") " pod="openstack/nova-scheduler-0" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.008349 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1317501c-8db9-4f03-a1cf-10e92b2ccc8b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1317501c-8db9-4f03-a1cf-10e92b2ccc8b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.009031 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1317501c-8db9-4f03-a1cf-10e92b2ccc8b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1317501c-8db9-4f03-a1cf-10e92b2ccc8b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.009195 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fd48f8c-e380-44cb-8c5b-8380b637e0ef-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7fd48f8c-e380-44cb-8c5b-8380b637e0ef\") " pod="openstack/nova-scheduler-0" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.009340 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/470abdb5-fefb-48c8-bb4c-9ad0e81caedb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"470abdb5-fefb-48c8-bb4c-9ad0e81caedb\") " pod="openstack/nova-metadata-0" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.009453 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ccbbp\" (UniqueName: \"kubernetes.io/projected/470abdb5-fefb-48c8-bb4c-9ad0e81caedb-kube-api-access-ccbbp\") pod \"nova-metadata-0\" (UID: \"470abdb5-fefb-48c8-bb4c-9ad0e81caedb\") " pod="openstack/nova-metadata-0" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.009556 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fd48f8c-e380-44cb-8c5b-8380b637e0ef-config-data\") pod \"nova-scheduler-0\" (UID: \"7fd48f8c-e380-44cb-8c5b-8380b637e0ef\") " pod="openstack/nova-scheduler-0" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.009713 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r4qbf\" (UniqueName: \"kubernetes.io/projected/1317501c-8db9-4f03-a1cf-10e92b2ccc8b-kube-api-access-r4qbf\") pod \"nova-cell1-novncproxy-0\" (UID: \"1317501c-8db9-4f03-a1cf-10e92b2ccc8b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.009866 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/470abdb5-fefb-48c8-bb4c-9ad0e81caedb-config-data\") pod \"nova-metadata-0\" (UID: \"470abdb5-fefb-48c8-bb4c-9ad0e81caedb\") " pod="openstack/nova-metadata-0" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.011733 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1317501c-8db9-4f03-a1cf-10e92b2ccc8b-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"1317501c-8db9-4f03-a1cf-10e92b2ccc8b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.014238 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-xbm92"] Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.022860 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-xbm92" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.024582 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1317501c-8db9-4f03-a1cf-10e92b2ccc8b-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"1317501c-8db9-4f03-a1cf-10e92b2ccc8b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.033706 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r4qbf\" (UniqueName: \"kubernetes.io/projected/1317501c-8db9-4f03-a1cf-10e92b2ccc8b-kube-api-access-r4qbf\") pod \"nova-cell1-novncproxy-0\" (UID: \"1317501c-8db9-4f03-a1cf-10e92b2ccc8b\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.039268 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-xbm92"] Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.100704 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.112026 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/470abdb5-fefb-48c8-bb4c-9ad0e81caedb-logs\") pod \"nova-metadata-0\" (UID: \"470abdb5-fefb-48c8-bb4c-9ad0e81caedb\") " pod="openstack/nova-metadata-0" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.112142 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a0b5b97-4605-4d4f-927f-d1eafe06956f-dns-svc\") pod \"dnsmasq-dns-bccf8f775-xbm92\" (UID: \"2a0b5b97-4605-4d4f-927f-d1eafe06956f\") " pod="openstack/dnsmasq-dns-bccf8f775-xbm92" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.112180 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ggr79\" (UniqueName: \"kubernetes.io/projected/7fd48f8c-e380-44cb-8c5b-8380b637e0ef-kube-api-access-ggr79\") pod \"nova-scheduler-0\" (UID: \"7fd48f8c-e380-44cb-8c5b-8380b637e0ef\") " pod="openstack/nova-scheduler-0" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.112228 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a0b5b97-4605-4d4f-927f-d1eafe06956f-config\") pod \"dnsmasq-dns-bccf8f775-xbm92\" (UID: \"2a0b5b97-4605-4d4f-927f-d1eafe06956f\") " pod="openstack/dnsmasq-dns-bccf8f775-xbm92" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.112257 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4rgrq\" (UniqueName: \"kubernetes.io/projected/2a0b5b97-4605-4d4f-927f-d1eafe06956f-kube-api-access-4rgrq\") pod \"dnsmasq-dns-bccf8f775-xbm92\" (UID: \"2a0b5b97-4605-4d4f-927f-d1eafe06956f\") " pod="openstack/dnsmasq-dns-bccf8f775-xbm92" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.112720 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fd48f8c-e380-44cb-8c5b-8380b637e0ef-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7fd48f8c-e380-44cb-8c5b-8380b637e0ef\") " pod="openstack/nova-scheduler-0" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.112780 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/470abdb5-fefb-48c8-bb4c-9ad0e81caedb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"470abdb5-fefb-48c8-bb4c-9ad0e81caedb\") " pod="openstack/nova-metadata-0" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.112819 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2a0b5b97-4605-4d4f-927f-d1eafe06956f-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-xbm92\" (UID: \"2a0b5b97-4605-4d4f-927f-d1eafe06956f\") " pod="openstack/dnsmasq-dns-bccf8f775-xbm92" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.112865 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ccbbp\" (UniqueName: \"kubernetes.io/projected/470abdb5-fefb-48c8-bb4c-9ad0e81caedb-kube-api-access-ccbbp\") pod \"nova-metadata-0\" (UID: \"470abdb5-fefb-48c8-bb4c-9ad0e81caedb\") " pod="openstack/nova-metadata-0" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.112911 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fd48f8c-e380-44cb-8c5b-8380b637e0ef-config-data\") pod \"nova-scheduler-0\" (UID: \"7fd48f8c-e380-44cb-8c5b-8380b637e0ef\") " pod="openstack/nova-scheduler-0" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.112935 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2a0b5b97-4605-4d4f-927f-d1eafe06956f-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-xbm92\" (UID: \"2a0b5b97-4605-4d4f-927f-d1eafe06956f\") " pod="openstack/dnsmasq-dns-bccf8f775-xbm92" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.113011 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2a0b5b97-4605-4d4f-927f-d1eafe06956f-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-xbm92\" (UID: \"2a0b5b97-4605-4d4f-927f-d1eafe06956f\") " pod="openstack/dnsmasq-dns-bccf8f775-xbm92" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.113089 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/470abdb5-fefb-48c8-bb4c-9ad0e81caedb-config-data\") pod \"nova-metadata-0\" (UID: \"470abdb5-fefb-48c8-bb4c-9ad0e81caedb\") " pod="openstack/nova-metadata-0" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.116402 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/470abdb5-fefb-48c8-bb4c-9ad0e81caedb-logs\") pod \"nova-metadata-0\" (UID: \"470abdb5-fefb-48c8-bb4c-9ad0e81caedb\") " pod="openstack/nova-metadata-0" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.118967 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.121559 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fd48f8c-e380-44cb-8c5b-8380b637e0ef-config-data\") pod \"nova-scheduler-0\" (UID: \"7fd48f8c-e380-44cb-8c5b-8380b637e0ef\") " pod="openstack/nova-scheduler-0" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.122663 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/470abdb5-fefb-48c8-bb4c-9ad0e81caedb-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"470abdb5-fefb-48c8-bb4c-9ad0e81caedb\") " pod="openstack/nova-metadata-0" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.126316 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/470abdb5-fefb-48c8-bb4c-9ad0e81caedb-config-data\") pod \"nova-metadata-0\" (UID: \"470abdb5-fefb-48c8-bb4c-9ad0e81caedb\") " pod="openstack/nova-metadata-0" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.130510 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fd48f8c-e380-44cb-8c5b-8380b637e0ef-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"7fd48f8c-e380-44cb-8c5b-8380b637e0ef\") " pod="openstack/nova-scheduler-0" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.132487 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ggr79\" (UniqueName: \"kubernetes.io/projected/7fd48f8c-e380-44cb-8c5b-8380b637e0ef-kube-api-access-ggr79\") pod \"nova-scheduler-0\" (UID: \"7fd48f8c-e380-44cb-8c5b-8380b637e0ef\") " pod="openstack/nova-scheduler-0" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.151659 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ccbbp\" (UniqueName: \"kubernetes.io/projected/470abdb5-fefb-48c8-bb4c-9ad0e81caedb-kube-api-access-ccbbp\") pod \"nova-metadata-0\" (UID: \"470abdb5-fefb-48c8-bb4c-9ad0e81caedb\") " pod="openstack/nova-metadata-0" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.217428 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a0b5b97-4605-4d4f-927f-d1eafe06956f-dns-svc\") pod \"dnsmasq-dns-bccf8f775-xbm92\" (UID: \"2a0b5b97-4605-4d4f-927f-d1eafe06956f\") " pod="openstack/dnsmasq-dns-bccf8f775-xbm92" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.217497 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a0b5b97-4605-4d4f-927f-d1eafe06956f-config\") pod \"dnsmasq-dns-bccf8f775-xbm92\" (UID: \"2a0b5b97-4605-4d4f-927f-d1eafe06956f\") " pod="openstack/dnsmasq-dns-bccf8f775-xbm92" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.219474 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a0b5b97-4605-4d4f-927f-d1eafe06956f-dns-svc\") pod \"dnsmasq-dns-bccf8f775-xbm92\" (UID: \"2a0b5b97-4605-4d4f-927f-d1eafe06956f\") " pod="openstack/dnsmasq-dns-bccf8f775-xbm92" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.220301 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a0b5b97-4605-4d4f-927f-d1eafe06956f-config\") pod \"dnsmasq-dns-bccf8f775-xbm92\" (UID: \"2a0b5b97-4605-4d4f-927f-d1eafe06956f\") " pod="openstack/dnsmasq-dns-bccf8f775-xbm92" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.220349 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4rgrq\" (UniqueName: \"kubernetes.io/projected/2a0b5b97-4605-4d4f-927f-d1eafe06956f-kube-api-access-4rgrq\") pod \"dnsmasq-dns-bccf8f775-xbm92\" (UID: \"2a0b5b97-4605-4d4f-927f-d1eafe06956f\") " pod="openstack/dnsmasq-dns-bccf8f775-xbm92" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.220458 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2a0b5b97-4605-4d4f-927f-d1eafe06956f-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-xbm92\" (UID: \"2a0b5b97-4605-4d4f-927f-d1eafe06956f\") " pod="openstack/dnsmasq-dns-bccf8f775-xbm92" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.220529 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2a0b5b97-4605-4d4f-927f-d1eafe06956f-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-xbm92\" (UID: \"2a0b5b97-4605-4d4f-927f-d1eafe06956f\") " pod="openstack/dnsmasq-dns-bccf8f775-xbm92" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.220589 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2a0b5b97-4605-4d4f-927f-d1eafe06956f-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-xbm92\" (UID: \"2a0b5b97-4605-4d4f-927f-d1eafe06956f\") " pod="openstack/dnsmasq-dns-bccf8f775-xbm92" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.221221 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2a0b5b97-4605-4d4f-927f-d1eafe06956f-ovsdbserver-sb\") pod \"dnsmasq-dns-bccf8f775-xbm92\" (UID: \"2a0b5b97-4605-4d4f-927f-d1eafe06956f\") " pod="openstack/dnsmasq-dns-bccf8f775-xbm92" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.221884 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2a0b5b97-4605-4d4f-927f-d1eafe06956f-ovsdbserver-nb\") pod \"dnsmasq-dns-bccf8f775-xbm92\" (UID: \"2a0b5b97-4605-4d4f-927f-d1eafe06956f\") " pod="openstack/dnsmasq-dns-bccf8f775-xbm92" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.222153 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2a0b5b97-4605-4d4f-927f-d1eafe06956f-dns-swift-storage-0\") pod \"dnsmasq-dns-bccf8f775-xbm92\" (UID: \"2a0b5b97-4605-4d4f-927f-d1eafe06956f\") " pod="openstack/dnsmasq-dns-bccf8f775-xbm92" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.238660 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4rgrq\" (UniqueName: \"kubernetes.io/projected/2a0b5b97-4605-4d4f-927f-d1eafe06956f-kube-api-access-4rgrq\") pod \"dnsmasq-dns-bccf8f775-xbm92\" (UID: \"2a0b5b97-4605-4d4f-927f-d1eafe06956f\") " pod="openstack/dnsmasq-dns-bccf8f775-xbm92" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.413160 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.429570 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.479168 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-xbm92" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.546987 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell0-cell-mapping-krtvp"] Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.639729 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e95b396-4e67-4df5-a39a-15ff3a00e6a1","Type":"ContainerStarted","Data":"d8cd6304732e32865c9d7f503d2a7316b66b5d1512696297666da84a07823c77"} Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.648783 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" event={"ID":"4390a950-7185-4fc0-b152-b5927940032b","Type":"ContainerStarted","Data":"cdd7251cfaa1a71292e8a6cac7e02004817db3e6278e1f8c2dd3ef78f445a274"} Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.649472 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.767109 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.827663 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dt5p8"] Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.828814 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-dt5p8" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.835348 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-scripts" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.835744 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.845403 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dt5p8"] Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.930559 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.949444 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/702714d0-d935-49ae-b19f-c0f80acd7758-scripts\") pod \"nova-cell1-conductor-db-sync-dt5p8\" (UID: \"702714d0-d935-49ae-b19f-c0f80acd7758\") " pod="openstack/nova-cell1-conductor-db-sync-dt5p8" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.949493 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/702714d0-d935-49ae-b19f-c0f80acd7758-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-dt5p8\" (UID: \"702714d0-d935-49ae-b19f-c0f80acd7758\") " pod="openstack/nova-cell1-conductor-db-sync-dt5p8" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.949573 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-phn76\" (UniqueName: \"kubernetes.io/projected/702714d0-d935-49ae-b19f-c0f80acd7758-kube-api-access-phn76\") pod \"nova-cell1-conductor-db-sync-dt5p8\" (UID: \"702714d0-d935-49ae-b19f-c0f80acd7758\") " pod="openstack/nova-cell1-conductor-db-sync-dt5p8" Oct 11 04:09:49 crc kubenswrapper[4967]: I1011 04:09:49.949594 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/702714d0-d935-49ae-b19f-c0f80acd7758-config-data\") pod \"nova-cell1-conductor-db-sync-dt5p8\" (UID: \"702714d0-d935-49ae-b19f-c0f80acd7758\") " pod="openstack/nova-cell1-conductor-db-sync-dt5p8" Oct 11 04:09:50 crc kubenswrapper[4967]: I1011 04:09:50.051443 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-phn76\" (UniqueName: \"kubernetes.io/projected/702714d0-d935-49ae-b19f-c0f80acd7758-kube-api-access-phn76\") pod \"nova-cell1-conductor-db-sync-dt5p8\" (UID: \"702714d0-d935-49ae-b19f-c0f80acd7758\") " pod="openstack/nova-cell1-conductor-db-sync-dt5p8" Oct 11 04:09:50 crc kubenswrapper[4967]: I1011 04:09:50.051500 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/702714d0-d935-49ae-b19f-c0f80acd7758-config-data\") pod \"nova-cell1-conductor-db-sync-dt5p8\" (UID: \"702714d0-d935-49ae-b19f-c0f80acd7758\") " pod="openstack/nova-cell1-conductor-db-sync-dt5p8" Oct 11 04:09:50 crc kubenswrapper[4967]: I1011 04:09:50.051614 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/702714d0-d935-49ae-b19f-c0f80acd7758-scripts\") pod \"nova-cell1-conductor-db-sync-dt5p8\" (UID: \"702714d0-d935-49ae-b19f-c0f80acd7758\") " pod="openstack/nova-cell1-conductor-db-sync-dt5p8" Oct 11 04:09:50 crc kubenswrapper[4967]: I1011 04:09:50.051673 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/702714d0-d935-49ae-b19f-c0f80acd7758-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-dt5p8\" (UID: \"702714d0-d935-49ae-b19f-c0f80acd7758\") " pod="openstack/nova-cell1-conductor-db-sync-dt5p8" Oct 11 04:09:50 crc kubenswrapper[4967]: I1011 04:09:50.057752 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/702714d0-d935-49ae-b19f-c0f80acd7758-combined-ca-bundle\") pod \"nova-cell1-conductor-db-sync-dt5p8\" (UID: \"702714d0-d935-49ae-b19f-c0f80acd7758\") " pod="openstack/nova-cell1-conductor-db-sync-dt5p8" Oct 11 04:09:50 crc kubenswrapper[4967]: I1011 04:09:50.062274 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/702714d0-d935-49ae-b19f-c0f80acd7758-scripts\") pod \"nova-cell1-conductor-db-sync-dt5p8\" (UID: \"702714d0-d935-49ae-b19f-c0f80acd7758\") " pod="openstack/nova-cell1-conductor-db-sync-dt5p8" Oct 11 04:09:50 crc kubenswrapper[4967]: I1011 04:09:50.062539 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/702714d0-d935-49ae-b19f-c0f80acd7758-config-data\") pod \"nova-cell1-conductor-db-sync-dt5p8\" (UID: \"702714d0-d935-49ae-b19f-c0f80acd7758\") " pod="openstack/nova-cell1-conductor-db-sync-dt5p8" Oct 11 04:09:50 crc kubenswrapper[4967]: I1011 04:09:50.074462 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-phn76\" (UniqueName: \"kubernetes.io/projected/702714d0-d935-49ae-b19f-c0f80acd7758-kube-api-access-phn76\") pod \"nova-cell1-conductor-db-sync-dt5p8\" (UID: \"702714d0-d935-49ae-b19f-c0f80acd7758\") " pod="openstack/nova-cell1-conductor-db-sync-dt5p8" Oct 11 04:09:50 crc kubenswrapper[4967]: I1011 04:09:50.176705 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 11 04:09:50 crc kubenswrapper[4967]: W1011 04:09:50.188770 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7fd48f8c_e380_44cb_8c5b_8380b637e0ef.slice/crio-2ae1533565367937fc2b5289cbea87e4c92e173b211be574ab405dfd4e8cdf7e WatchSource:0}: Error finding container 2ae1533565367937fc2b5289cbea87e4c92e173b211be574ab405dfd4e8cdf7e: Status 404 returned error can't find the container with id 2ae1533565367937fc2b5289cbea87e4c92e173b211be574ab405dfd4e8cdf7e Oct 11 04:09:50 crc kubenswrapper[4967]: I1011 04:09:50.228656 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-dt5p8" Oct 11 04:09:50 crc kubenswrapper[4967]: I1011 04:09:50.369722 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-xbm92"] Oct 11 04:09:50 crc kubenswrapper[4967]: I1011 04:09:50.390251 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 11 04:09:50 crc kubenswrapper[4967]: W1011 04:09:50.399379 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2a0b5b97_4605_4d4f_927f_d1eafe06956f.slice/crio-38de4d739ca79b2d47266353bb6177bd566656f98f4c935c1653a61487cd1e25 WatchSource:0}: Error finding container 38de4d739ca79b2d47266353bb6177bd566656f98f4c935c1653a61487cd1e25: Status 404 returned error can't find the container with id 38de4d739ca79b2d47266353bb6177bd566656f98f4c935c1653a61487cd1e25 Oct 11 04:09:50 crc kubenswrapper[4967]: I1011 04:09:50.715443 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"470abdb5-fefb-48c8-bb4c-9ad0e81caedb","Type":"ContainerStarted","Data":"7a9539f736ba9147c2dffe84ce3e68562bedbcbbdddb76b85bc380fd37718743"} Oct 11 04:09:50 crc kubenswrapper[4967]: I1011 04:09:50.758159 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-krtvp" event={"ID":"9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e","Type":"ContainerStarted","Data":"4bbf287d1de495b499f6b4194a0015fc0ee50f432089740f8c0298235e83628f"} Oct 11 04:09:50 crc kubenswrapper[4967]: I1011 04:09:50.758203 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-krtvp" event={"ID":"9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e","Type":"ContainerStarted","Data":"abe385c938c733b5667d29545ace1b41a2dd946bdab41adba0750f742c0c9fa6"} Oct 11 04:09:50 crc kubenswrapper[4967]: I1011 04:09:50.760353 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7fd48f8c-e380-44cb-8c5b-8380b637e0ef","Type":"ContainerStarted","Data":"2ae1533565367937fc2b5289cbea87e4c92e173b211be574ab405dfd4e8cdf7e"} Oct 11 04:09:50 crc kubenswrapper[4967]: I1011 04:09:50.783546 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-xbm92" event={"ID":"2a0b5b97-4605-4d4f-927f-d1eafe06956f","Type":"ContainerStarted","Data":"38de4d739ca79b2d47266353bb6177bd566656f98f4c935c1653a61487cd1e25"} Oct 11 04:09:50 crc kubenswrapper[4967]: I1011 04:09:50.792435 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell0-cell-mapping-krtvp" podStartSLOduration=2.792416251 podStartE2EDuration="2.792416251s" podCreationTimestamp="2025-10-11 04:09:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:09:50.782494311 +0000 UTC m=+1118.745703244" watchObservedRunningTime="2025-10-11 04:09:50.792416251 +0000 UTC m=+1118.755625184" Oct 11 04:09:50 crc kubenswrapper[4967]: I1011 04:09:50.802986 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cee26db2-5786-43ed-8a26-3b0e55ca2964","Type":"ContainerStarted","Data":"e1e2cff23ec999a1f5978cc9669d532364f773752e4c1b9d3bb294a9796aa86f"} Oct 11 04:09:50 crc kubenswrapper[4967]: I1011 04:09:50.901610 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1317501c-8db9-4f03-a1cf-10e92b2ccc8b","Type":"ContainerStarted","Data":"9240fd6ade19d6577d0fda15b5465bcf854a613fc755c893f4a325590d9a1523"} Oct 11 04:09:50 crc kubenswrapper[4967]: I1011 04:09:50.901693 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 11 04:09:50 crc kubenswrapper[4967]: I1011 04:09:50.910331 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=2.2342912950000002 podStartE2EDuration="6.910305771s" podCreationTimestamp="2025-10-11 04:09:44 +0000 UTC" firstStartedPulling="2025-10-11 04:09:45.484258986 +0000 UTC m=+1113.447467929" lastFinishedPulling="2025-10-11 04:09:50.160273472 +0000 UTC m=+1118.123482405" observedRunningTime="2025-10-11 04:09:50.879140129 +0000 UTC m=+1118.842349062" watchObservedRunningTime="2025-10-11 04:09:50.910305771 +0000 UTC m=+1118.873514704" Oct 11 04:09:50 crc kubenswrapper[4967]: I1011 04:09:50.972646 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dt5p8"] Oct 11 04:09:50 crc kubenswrapper[4967]: W1011 04:09:50.998197 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod702714d0_d935_49ae_b19f_c0f80acd7758.slice/crio-7e6407148046234336abdd1c59ec23032385988bf042f14b941198b1e42c9317 WatchSource:0}: Error finding container 7e6407148046234336abdd1c59ec23032385988bf042f14b941198b1e42c9317: Status 404 returned error can't find the container with id 7e6407148046234336abdd1c59ec23032385988bf042f14b941198b1e42c9317 Oct 11 04:09:51 crc kubenswrapper[4967]: I1011 04:09:51.901709 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-dt5p8" event={"ID":"702714d0-d935-49ae-b19f-c0f80acd7758","Type":"ContainerStarted","Data":"8bdf9296be5a3e6cedf0cf960a89608cbc5c90a2175a407f8d332f1f82d80cc8"} Oct 11 04:09:51 crc kubenswrapper[4967]: I1011 04:09:51.902123 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-dt5p8" event={"ID":"702714d0-d935-49ae-b19f-c0f80acd7758","Type":"ContainerStarted","Data":"7e6407148046234336abdd1c59ec23032385988bf042f14b941198b1e42c9317"} Oct 11 04:09:51 crc kubenswrapper[4967]: I1011 04:09:51.908824 4967 generic.go:334] "Generic (PLEG): container finished" podID="2a0b5b97-4605-4d4f-927f-d1eafe06956f" containerID="050e060ac770a18c635b538f899e5289cf36e4d693568ba33a0b57bd48d0eb85" exitCode=0 Oct 11 04:09:51 crc kubenswrapper[4967]: I1011 04:09:51.908891 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-xbm92" event={"ID":"2a0b5b97-4605-4d4f-927f-d1eafe06956f","Type":"ContainerDied","Data":"050e060ac770a18c635b538f899e5289cf36e4d693568ba33a0b57bd48d0eb85"} Oct 11 04:09:51 crc kubenswrapper[4967]: I1011 04:09:51.914325 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e95b396-4e67-4df5-a39a-15ff3a00e6a1","Type":"ContainerStarted","Data":"7da08d64f43a6586cbd116ace09f566fd2c959ac5ecddb850663fc0be41dc2f2"} Oct 11 04:09:51 crc kubenswrapper[4967]: I1011 04:09:51.928309 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-db-sync-dt5p8" podStartSLOduration=2.9282880049999997 podStartE2EDuration="2.928288005s" podCreationTimestamp="2025-10-11 04:09:49 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:09:51.918622741 +0000 UTC m=+1119.881831674" watchObservedRunningTime="2025-10-11 04:09:51.928288005 +0000 UTC m=+1119.891496938" Oct 11 04:09:52 crc kubenswrapper[4967]: I1011 04:09:52.795904 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 11 04:09:52 crc kubenswrapper[4967]: I1011 04:09:52.829247 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 11 04:09:52 crc kubenswrapper[4967]: I1011 04:09:52.947398 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-xbm92" event={"ID":"2a0b5b97-4605-4d4f-927f-d1eafe06956f","Type":"ContainerStarted","Data":"e42a65cb81472bcfcb031eb8b1298f5038365ed72c88b70eac57614ce427dde4"} Oct 11 04:09:52 crc kubenswrapper[4967]: I1011 04:09:52.947614 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-bccf8f775-xbm92" Oct 11 04:09:52 crc kubenswrapper[4967]: I1011 04:09:52.971332 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-bccf8f775-xbm92" podStartSLOduration=4.971313139 podStartE2EDuration="4.971313139s" podCreationTimestamp="2025-10-11 04:09:48 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:09:52.966813215 +0000 UTC m=+1120.930022148" watchObservedRunningTime="2025-10-11 04:09:52.971313139 +0000 UTC m=+1120.934522072" Oct 11 04:09:54 crc kubenswrapper[4967]: I1011 04:09:54.476178 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-neutron-agent-657f6548d9-9fb8w" Oct 11 04:09:59 crc kubenswrapper[4967]: I1011 04:09:59.000327 4967 generic.go:334] "Generic (PLEG): container finished" podID="9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e" containerID="4bbf287d1de495b499f6b4194a0015fc0ee50f432089740f8c0298235e83628f" exitCode=0 Oct 11 04:09:59 crc kubenswrapper[4967]: I1011 04:09:59.000771 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-krtvp" event={"ID":"9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e","Type":"ContainerDied","Data":"4bbf287d1de495b499f6b4194a0015fc0ee50f432089740f8c0298235e83628f"} Oct 11 04:09:59 crc kubenswrapper[4967]: I1011 04:09:59.481285 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-bccf8f775-xbm92" Oct 11 04:09:59 crc kubenswrapper[4967]: I1011 04:09:59.559640 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-vhtnn"] Oct 11 04:09:59 crc kubenswrapper[4967]: I1011 04:09:59.559967 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-6578955fd5-vhtnn" podUID="60a1ad86-dc2a-46a3-8c63-6e48d3a2d291" containerName="dnsmasq-dns" containerID="cri-o://f98f49cc939cb150c97d159f63d67e4ee4bfc1f905784b0d426a80a71c4386a3" gracePeriod=10 Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.011578 4967 generic.go:334] "Generic (PLEG): container finished" podID="60a1ad86-dc2a-46a3-8c63-6e48d3a2d291" containerID="f98f49cc939cb150c97d159f63d67e4ee4bfc1f905784b0d426a80a71c4386a3" exitCode=0 Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.012231 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-vhtnn" event={"ID":"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291","Type":"ContainerDied","Data":"f98f49cc939cb150c97d159f63d67e4ee4bfc1f905784b0d426a80a71c4386a3"} Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.014588 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"470abdb5-fefb-48c8-bb4c-9ad0e81caedb","Type":"ContainerStarted","Data":"19de676933f0fbabbd00149485dff306db57a2781282138364b5f63f2e7a4a70"} Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.017608 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"64c75ea6-aed6-4bf6-8eef-39e76384e14f","Type":"ContainerStarted","Data":"93d7b8ee4a62c1ecfe52437af2caf73381136995aee5a103e35e85022d1f814a"} Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.041391 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7fd48f8c-e380-44cb-8c5b-8380b637e0ef","Type":"ContainerStarted","Data":"4726e4a657dc964b9f92f99fa450dccebcb517bf23d903b7075ee3e034e8ccd1"} Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.048946 4967 generic.go:334] "Generic (PLEG): container finished" podID="95c0b2ec-d176-4832-897b-7c33176243cd" containerID="72c44f733e48d67f733027b1618bd909b98791f2541d0ea0dc85aaee79055f6e" exitCode=0 Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.049019 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"95c0b2ec-d176-4832-897b-7c33176243cd","Type":"ContainerDied","Data":"72c44f733e48d67f733027b1618bd909b98791f2541d0ea0dc85aaee79055f6e"} Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.062347 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cee26db2-5786-43ed-8a26-3b0e55ca2964","Type":"ContainerStarted","Data":"409c9974b4f4831aab22106bce09e00932e188a272394ae0cf2cd9a2dcaa84f4"} Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.082303 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-cell1-novncproxy-0" podUID="1317501c-8db9-4f03-a1cf-10e92b2ccc8b" containerName="nova-cell1-novncproxy-novncproxy" containerID="cri-o://fb17607f6184c47fb87463585b50b5bac67a191f9c083416e9f65b664822e600" gracePeriod=30 Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.082399 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1317501c-8db9-4f03-a1cf-10e92b2ccc8b","Type":"ContainerStarted","Data":"fb17607f6184c47fb87463585b50b5bac67a191f9c083416e9f65b664822e600"} Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.112957 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.036060073 podStartE2EDuration="12.112931282s" podCreationTimestamp="2025-10-11 04:09:48 +0000 UTC" firstStartedPulling="2025-10-11 04:09:50.211316874 +0000 UTC m=+1118.174525807" lastFinishedPulling="2025-10-11 04:09:59.288188083 +0000 UTC m=+1127.251397016" observedRunningTime="2025-10-11 04:10:00.104913116 +0000 UTC m=+1128.068122069" watchObservedRunningTime="2025-10-11 04:10:00.112931282 +0000 UTC m=+1128.076140235" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.126324 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-vhtnn" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.195865 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.763203324 podStartE2EDuration="12.195848342s" podCreationTimestamp="2025-10-11 04:09:48 +0000 UTC" firstStartedPulling="2025-10-11 04:09:49.841559801 +0000 UTC m=+1117.804768734" lastFinishedPulling="2025-10-11 04:09:59.274204819 +0000 UTC m=+1127.237413752" observedRunningTime="2025-10-11 04:10:00.151315831 +0000 UTC m=+1128.114524774" watchObservedRunningTime="2025-10-11 04:10:00.195848342 +0000 UTC m=+1128.159057265" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.205585 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.83124003 podStartE2EDuration="12.205544907s" podCreationTimestamp="2025-10-11 04:09:48 +0000 UTC" firstStartedPulling="2025-10-11 04:09:49.938911006 +0000 UTC m=+1117.902119939" lastFinishedPulling="2025-10-11 04:09:59.313215883 +0000 UTC m=+1127.276424816" observedRunningTime="2025-10-11 04:10:00.170529356 +0000 UTC m=+1128.133738299" watchObservedRunningTime="2025-10-11 04:10:00.205544907 +0000 UTC m=+1128.168753840" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.295834 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z56rs\" (UniqueName: \"kubernetes.io/projected/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-kube-api-access-z56rs\") pod \"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291\" (UID: \"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291\") " Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.295923 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-config\") pod \"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291\" (UID: \"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291\") " Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.295952 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-ovsdbserver-sb\") pod \"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291\" (UID: \"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291\") " Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.296015 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-ovsdbserver-nb\") pod \"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291\" (UID: \"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291\") " Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.296041 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-dns-swift-storage-0\") pod \"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291\" (UID: \"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291\") " Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.296144 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-dns-svc\") pod \"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291\" (UID: \"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291\") " Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.301394 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-kube-api-access-z56rs" (OuterVolumeSpecName: "kube-api-access-z56rs") pod "60a1ad86-dc2a-46a3-8c63-6e48d3a2d291" (UID: "60a1ad86-dc2a-46a3-8c63-6e48d3a2d291"). InnerVolumeSpecName "kube-api-access-z56rs". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.373590 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "60a1ad86-dc2a-46a3-8c63-6e48d3a2d291" (UID: "60a1ad86-dc2a-46a3-8c63-6e48d3a2d291"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.386489 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-config" (OuterVolumeSpecName: "config") pod "60a1ad86-dc2a-46a3-8c63-6e48d3a2d291" (UID: "60a1ad86-dc2a-46a3-8c63-6e48d3a2d291"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.397524 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "60a1ad86-dc2a-46a3-8c63-6e48d3a2d291" (UID: "60a1ad86-dc2a-46a3-8c63-6e48d3a2d291"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.400795 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z56rs\" (UniqueName: \"kubernetes.io/projected/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-kube-api-access-z56rs\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.400826 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-config\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.400836 4967 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.400844 4967 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.402969 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "60a1ad86-dc2a-46a3-8c63-6e48d3a2d291" (UID: "60a1ad86-dc2a-46a3-8c63-6e48d3a2d291"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.405495 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "60a1ad86-dc2a-46a3-8c63-6e48d3a2d291" (UID: "60a1ad86-dc2a-46a3-8c63-6e48d3a2d291"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.502985 4967 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.503016 4967 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.655919 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-krtvp" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.679104 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-0" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.819431 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95c0b2ec-d176-4832-897b-7c33176243cd-scripts\") pod \"95c0b2ec-d176-4832-897b-7c33176243cd\" (UID: \"95c0b2ec-d176-4832-897b-7c33176243cd\") " Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.819514 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e-combined-ca-bundle\") pod \"9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e\" (UID: \"9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e\") " Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.819564 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-scd22\" (UniqueName: \"kubernetes.io/projected/9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e-kube-api-access-scd22\") pod \"9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e\" (UID: \"9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e\") " Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.819637 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e-scripts\") pod \"9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e\" (UID: \"9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e\") " Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.819690 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/95c0b2ec-d176-4832-897b-7c33176243cd-var-lib-ironic\") pod \"95c0b2ec-d176-4832-897b-7c33176243cd\" (UID: \"95c0b2ec-d176-4832-897b-7c33176243cd\") " Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.819718 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e-config-data\") pod \"9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e\" (UID: \"9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e\") " Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.819746 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/95c0b2ec-d176-4832-897b-7c33176243cd-etc-podinfo\") pod \"95c0b2ec-d176-4832-897b-7c33176243cd\" (UID: \"95c0b2ec-d176-4832-897b-7c33176243cd\") " Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.819774 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rbmng\" (UniqueName: \"kubernetes.io/projected/95c0b2ec-d176-4832-897b-7c33176243cd-kube-api-access-rbmng\") pod \"95c0b2ec-d176-4832-897b-7c33176243cd\" (UID: \"95c0b2ec-d176-4832-897b-7c33176243cd\") " Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.819800 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/95c0b2ec-d176-4832-897b-7c33176243cd-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"95c0b2ec-d176-4832-897b-7c33176243cd\" (UID: \"95c0b2ec-d176-4832-897b-7c33176243cd\") " Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.819838 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95c0b2ec-d176-4832-897b-7c33176243cd-combined-ca-bundle\") pod \"95c0b2ec-d176-4832-897b-7c33176243cd\" (UID: \"95c0b2ec-d176-4832-897b-7c33176243cd\") " Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.819887 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/95c0b2ec-d176-4832-897b-7c33176243cd-config\") pod \"95c0b2ec-d176-4832-897b-7c33176243cd\" (UID: \"95c0b2ec-d176-4832-897b-7c33176243cd\") " Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.820614 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95c0b2ec-d176-4832-897b-7c33176243cd-var-lib-ironic" (OuterVolumeSpecName: "var-lib-ironic") pod "95c0b2ec-d176-4832-897b-7c33176243cd" (UID: "95c0b2ec-d176-4832-897b-7c33176243cd"). InnerVolumeSpecName "var-lib-ironic". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.820892 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/95c0b2ec-d176-4832-897b-7c33176243cd-var-lib-ironic-inspector-dhcp-hostsdir" (OuterVolumeSpecName: "var-lib-ironic-inspector-dhcp-hostsdir") pod "95c0b2ec-d176-4832-897b-7c33176243cd" (UID: "95c0b2ec-d176-4832-897b-7c33176243cd"). InnerVolumeSpecName "var-lib-ironic-inspector-dhcp-hostsdir". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.824717 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95c0b2ec-d176-4832-897b-7c33176243cd-config" (OuterVolumeSpecName: "config") pod "95c0b2ec-d176-4832-897b-7c33176243cd" (UID: "95c0b2ec-d176-4832-897b-7c33176243cd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.825426 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/95c0b2ec-d176-4832-897b-7c33176243cd-kube-api-access-rbmng" (OuterVolumeSpecName: "kube-api-access-rbmng") pod "95c0b2ec-d176-4832-897b-7c33176243cd" (UID: "95c0b2ec-d176-4832-897b-7c33176243cd"). InnerVolumeSpecName "kube-api-access-rbmng". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.827507 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95c0b2ec-d176-4832-897b-7c33176243cd-scripts" (OuterVolumeSpecName: "scripts") pod "95c0b2ec-d176-4832-897b-7c33176243cd" (UID: "95c0b2ec-d176-4832-897b-7c33176243cd"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.827574 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e-kube-api-access-scd22" (OuterVolumeSpecName: "kube-api-access-scd22") pod "9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e" (UID: "9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e"). InnerVolumeSpecName "kube-api-access-scd22". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.828413 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e-scripts" (OuterVolumeSpecName: "scripts") pod "9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e" (UID: "9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.830598 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/95c0b2ec-d176-4832-897b-7c33176243cd-etc-podinfo" (OuterVolumeSpecName: "etc-podinfo") pod "95c0b2ec-d176-4832-897b-7c33176243cd" (UID: "95c0b2ec-d176-4832-897b-7c33176243cd"). InnerVolumeSpecName "etc-podinfo". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.853761 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e-config-data" (OuterVolumeSpecName: "config-data") pod "9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e" (UID: "9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.879317 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e" (UID: "9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.892658 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/95c0b2ec-d176-4832-897b-7c33176243cd-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "95c0b2ec-d176-4832-897b-7c33176243cd" (UID: "95c0b2ec-d176-4832-897b-7c33176243cd"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.922325 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.922359 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-scd22\" (UniqueName: \"kubernetes.io/projected/9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e-kube-api-access-scd22\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.922370 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.922381 4967 reconciler_common.go:293] "Volume detached for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/95c0b2ec-d176-4832-897b-7c33176243cd-var-lib-ironic\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.922391 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.922400 4967 reconciler_common.go:293] "Volume detached for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/95c0b2ec-d176-4832-897b-7c33176243cd-etc-podinfo\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.922409 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rbmng\" (UniqueName: \"kubernetes.io/projected/95c0b2ec-d176-4832-897b-7c33176243cd-kube-api-access-rbmng\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.922420 4967 reconciler_common.go:293] "Volume detached for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/95c0b2ec-d176-4832-897b-7c33176243cd-var-lib-ironic-inspector-dhcp-hostsdir\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.922428 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/95c0b2ec-d176-4832-897b-7c33176243cd-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.922439 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/secret/95c0b2ec-d176-4832-897b-7c33176243cd-config\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:00 crc kubenswrapper[4967]: I1011 04:10:00.922449 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/95c0b2ec-d176-4832-897b-7c33176243cd-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.095381 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-6578955fd5-vhtnn" event={"ID":"60a1ad86-dc2a-46a3-8c63-6e48d3a2d291","Type":"ContainerDied","Data":"1e99736afe34e37994f390afffddc363850018f5aaeb843222ffc2ba17efac02"} Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.095750 4967 scope.go:117] "RemoveContainer" containerID="f98f49cc939cb150c97d159f63d67e4ee4bfc1f905784b0d426a80a71c4386a3" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.095626 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-6578955fd5-vhtnn" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.099652 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"470abdb5-fefb-48c8-bb4c-9ad0e81caedb","Type":"ContainerStarted","Data":"5584621212aa51c0e1c6bdac7220d1190ccbbf6c7c75ee751ac70c2bce5cf42d"} Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.099819 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="470abdb5-fefb-48c8-bb4c-9ad0e81caedb" containerName="nova-metadata-log" containerID="cri-o://19de676933f0fbabbd00149485dff306db57a2781282138364b5f63f2e7a4a70" gracePeriod=30 Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.099948 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="470abdb5-fefb-48c8-bb4c-9ad0e81caedb" containerName="nova-metadata-metadata" containerID="cri-o://5584621212aa51c0e1c6bdac7220d1190ccbbf6c7c75ee751ac70c2bce5cf42d" gracePeriod=30 Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.106392 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell0-cell-mapping-krtvp" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.106945 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell0-cell-mapping-krtvp" event={"ID":"9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e","Type":"ContainerDied","Data":"abe385c938c733b5667d29545ace1b41a2dd946bdab41adba0750f742c0c9fa6"} Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.106992 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="abe385c938c733b5667d29545ace1b41a2dd946bdab41adba0750f742c0c9fa6" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.115444 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"95c0b2ec-d176-4832-897b-7c33176243cd","Type":"ContainerDied","Data":"cb8b484056c96f18473c916e31fbf8c80627fb448096b792c38c4ef2b2bb6e72"} Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.115480 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-0" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.121642 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cee26db2-5786-43ed-8a26-3b0e55ca2964","Type":"ContainerStarted","Data":"aba70ed5fba8bf5f924ca4861b46c22417dbe426f6ec5a6b879cb253cdece0ef"} Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.139022 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-vhtnn"] Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.141156 4967 scope.go:117] "RemoveContainer" containerID="da1a30c39594c793da68e5ba312b7e700b454828fae3142a32c3eddfaaa5e815" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.151719 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-6578955fd5-vhtnn"] Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.158700 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=4.283613815 podStartE2EDuration="13.1586865s" podCreationTimestamp="2025-10-11 04:09:48 +0000 UTC" firstStartedPulling="2025-10-11 04:09:50.437712157 +0000 UTC m=+1118.400921090" lastFinishedPulling="2025-10-11 04:09:59.312784842 +0000 UTC m=+1127.275993775" observedRunningTime="2025-10-11 04:10:01.145768071 +0000 UTC m=+1129.108977004" watchObservedRunningTime="2025-10-11 04:10:01.1586865 +0000 UTC m=+1129.121895433" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.188086 4967 scope.go:117] "RemoveContainer" containerID="72c44f733e48d67f733027b1618bd909b98791f2541d0ea0dc85aaee79055f6e" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.297989 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.321186 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-inspector-0"] Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.336662 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ironic-inspector-0"] Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.345341 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.360646 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ironic-inspector-0"] Oct 11 04:10:01 crc kubenswrapper[4967]: E1011 04:10:01.361102 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e" containerName="nova-manage" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.361120 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e" containerName="nova-manage" Oct 11 04:10:01 crc kubenswrapper[4967]: E1011 04:10:01.361141 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60a1ad86-dc2a-46a3-8c63-6e48d3a2d291" containerName="dnsmasq-dns" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.361149 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="60a1ad86-dc2a-46a3-8c63-6e48d3a2d291" containerName="dnsmasq-dns" Oct 11 04:10:01 crc kubenswrapper[4967]: E1011 04:10:01.361166 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="95c0b2ec-d176-4832-897b-7c33176243cd" containerName="ironic-python-agent-init" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.361171 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="95c0b2ec-d176-4832-897b-7c33176243cd" containerName="ironic-python-agent-init" Oct 11 04:10:01 crc kubenswrapper[4967]: E1011 04:10:01.361197 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60a1ad86-dc2a-46a3-8c63-6e48d3a2d291" containerName="init" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.361204 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="60a1ad86-dc2a-46a3-8c63-6e48d3a2d291" containerName="init" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.361378 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="60a1ad86-dc2a-46a3-8c63-6e48d3a2d291" containerName="dnsmasq-dns" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.361399 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e" containerName="nova-manage" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.361410 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="95c0b2ec-d176-4832-897b-7c33176243cd" containerName="ironic-python-agent-init" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.363877 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-0" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.369698 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-0"] Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.373959 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-scripts" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.373981 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ironic-inspector-config-data" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.374230 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ironic-inspector-internal-svc" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.374263 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ironic-inspector-public-svc" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.546060 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c619427d-62f7-463c-8262-d3b9fdd6ade0-scripts\") pod \"ironic-inspector-0\" (UID: \"c619427d-62f7-463c-8262-d3b9fdd6ade0\") " pod="openstack/ironic-inspector-0" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.546156 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/c619427d-62f7-463c-8262-d3b9fdd6ade0-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-0\" (UID: \"c619427d-62f7-463c-8262-d3b9fdd6ade0\") " pod="openstack/ironic-inspector-0" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.546185 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c619427d-62f7-463c-8262-d3b9fdd6ade0-public-tls-certs\") pod \"ironic-inspector-0\" (UID: \"c619427d-62f7-463c-8262-d3b9fdd6ade0\") " pod="openstack/ironic-inspector-0" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.546230 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/c619427d-62f7-463c-8262-d3b9fdd6ade0-var-lib-ironic\") pod \"ironic-inspector-0\" (UID: \"c619427d-62f7-463c-8262-d3b9fdd6ade0\") " pod="openstack/ironic-inspector-0" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.546351 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5p9g\" (UniqueName: \"kubernetes.io/projected/c619427d-62f7-463c-8262-d3b9fdd6ade0-kube-api-access-r5p9g\") pod \"ironic-inspector-0\" (UID: \"c619427d-62f7-463c-8262-d3b9fdd6ade0\") " pod="openstack/ironic-inspector-0" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.546419 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c619427d-62f7-463c-8262-d3b9fdd6ade0-internal-tls-certs\") pod \"ironic-inspector-0\" (UID: \"c619427d-62f7-463c-8262-d3b9fdd6ade0\") " pod="openstack/ironic-inspector-0" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.546505 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c619427d-62f7-463c-8262-d3b9fdd6ade0-combined-ca-bundle\") pod \"ironic-inspector-0\" (UID: \"c619427d-62f7-463c-8262-d3b9fdd6ade0\") " pod="openstack/ironic-inspector-0" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.546623 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/c619427d-62f7-463c-8262-d3b9fdd6ade0-etc-podinfo\") pod \"ironic-inspector-0\" (UID: \"c619427d-62f7-463c-8262-d3b9fdd6ade0\") " pod="openstack/ironic-inspector-0" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.546657 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c619427d-62f7-463c-8262-d3b9fdd6ade0-config\") pod \"ironic-inspector-0\" (UID: \"c619427d-62f7-463c-8262-d3b9fdd6ade0\") " pod="openstack/ironic-inspector-0" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.647875 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/c619427d-62f7-463c-8262-d3b9fdd6ade0-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-0\" (UID: \"c619427d-62f7-463c-8262-d3b9fdd6ade0\") " pod="openstack/ironic-inspector-0" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.647926 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c619427d-62f7-463c-8262-d3b9fdd6ade0-public-tls-certs\") pod \"ironic-inspector-0\" (UID: \"c619427d-62f7-463c-8262-d3b9fdd6ade0\") " pod="openstack/ironic-inspector-0" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.647974 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/c619427d-62f7-463c-8262-d3b9fdd6ade0-var-lib-ironic\") pod \"ironic-inspector-0\" (UID: \"c619427d-62f7-463c-8262-d3b9fdd6ade0\") " pod="openstack/ironic-inspector-0" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.648005 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5p9g\" (UniqueName: \"kubernetes.io/projected/c619427d-62f7-463c-8262-d3b9fdd6ade0-kube-api-access-r5p9g\") pod \"ironic-inspector-0\" (UID: \"c619427d-62f7-463c-8262-d3b9fdd6ade0\") " pod="openstack/ironic-inspector-0" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.648034 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c619427d-62f7-463c-8262-d3b9fdd6ade0-internal-tls-certs\") pod \"ironic-inspector-0\" (UID: \"c619427d-62f7-463c-8262-d3b9fdd6ade0\") " pod="openstack/ironic-inspector-0" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.648088 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c619427d-62f7-463c-8262-d3b9fdd6ade0-combined-ca-bundle\") pod \"ironic-inspector-0\" (UID: \"c619427d-62f7-463c-8262-d3b9fdd6ade0\") " pod="openstack/ironic-inspector-0" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.648136 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/c619427d-62f7-463c-8262-d3b9fdd6ade0-etc-podinfo\") pod \"ironic-inspector-0\" (UID: \"c619427d-62f7-463c-8262-d3b9fdd6ade0\") " pod="openstack/ironic-inspector-0" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.648160 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/secret/c619427d-62f7-463c-8262-d3b9fdd6ade0-config\") pod \"ironic-inspector-0\" (UID: \"c619427d-62f7-463c-8262-d3b9fdd6ade0\") " pod="openstack/ironic-inspector-0" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.648182 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c619427d-62f7-463c-8262-d3b9fdd6ade0-scripts\") pod \"ironic-inspector-0\" (UID: \"c619427d-62f7-463c-8262-d3b9fdd6ade0\") " pod="openstack/ironic-inspector-0" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.650735 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-ironic-inspector-dhcp-hostsdir\" (UniqueName: \"kubernetes.io/empty-dir/c619427d-62f7-463c-8262-d3b9fdd6ade0-var-lib-ironic-inspector-dhcp-hostsdir\") pod \"ironic-inspector-0\" (UID: \"c619427d-62f7-463c-8262-d3b9fdd6ade0\") " pod="openstack/ironic-inspector-0" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.653386 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-ironic\" (UniqueName: \"kubernetes.io/empty-dir/c619427d-62f7-463c-8262-d3b9fdd6ade0-var-lib-ironic\") pod \"ironic-inspector-0\" (UID: \"c619427d-62f7-463c-8262-d3b9fdd6ade0\") " pod="openstack/ironic-inspector-0" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.664205 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c619427d-62f7-463c-8262-d3b9fdd6ade0-scripts\") pod \"ironic-inspector-0\" (UID: \"c619427d-62f7-463c-8262-d3b9fdd6ade0\") " pod="openstack/ironic-inspector-0" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.671611 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-podinfo\" (UniqueName: \"kubernetes.io/downward-api/c619427d-62f7-463c-8262-d3b9fdd6ade0-etc-podinfo\") pod \"ironic-inspector-0\" (UID: \"c619427d-62f7-463c-8262-d3b9fdd6ade0\") " pod="openstack/ironic-inspector-0" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.672146 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c619427d-62f7-463c-8262-d3b9fdd6ade0-combined-ca-bundle\") pod \"ironic-inspector-0\" (UID: \"c619427d-62f7-463c-8262-d3b9fdd6ade0\") " pod="openstack/ironic-inspector-0" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.672828 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/c619427d-62f7-463c-8262-d3b9fdd6ade0-public-tls-certs\") pod \"ironic-inspector-0\" (UID: \"c619427d-62f7-463c-8262-d3b9fdd6ade0\") " pod="openstack/ironic-inspector-0" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.673360 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/c619427d-62f7-463c-8262-d3b9fdd6ade0-internal-tls-certs\") pod \"ironic-inspector-0\" (UID: \"c619427d-62f7-463c-8262-d3b9fdd6ade0\") " pod="openstack/ironic-inspector-0" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.673404 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/secret/c619427d-62f7-463c-8262-d3b9fdd6ade0-config\") pod \"ironic-inspector-0\" (UID: \"c619427d-62f7-463c-8262-d3b9fdd6ade0\") " pod="openstack/ironic-inspector-0" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.677791 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5p9g\" (UniqueName: \"kubernetes.io/projected/c619427d-62f7-463c-8262-d3b9fdd6ade0-kube-api-access-r5p9g\") pod \"ironic-inspector-0\" (UID: \"c619427d-62f7-463c-8262-d3b9fdd6ade0\") " pod="openstack/ironic-inspector-0" Oct 11 04:10:01 crc kubenswrapper[4967]: I1011 04:10:01.724033 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ironic-inspector-0" Oct 11 04:10:02 crc kubenswrapper[4967]: I1011 04:10:02.131342 4967 generic.go:334] "Generic (PLEG): container finished" podID="470abdb5-fefb-48c8-bb4c-9ad0e81caedb" containerID="5584621212aa51c0e1c6bdac7220d1190ccbbf6c7c75ee751ac70c2bce5cf42d" exitCode=0 Oct 11 04:10:02 crc kubenswrapper[4967]: I1011 04:10:02.131648 4967 generic.go:334] "Generic (PLEG): container finished" podID="470abdb5-fefb-48c8-bb4c-9ad0e81caedb" containerID="19de676933f0fbabbd00149485dff306db57a2781282138364b5f63f2e7a4a70" exitCode=143 Oct 11 04:10:02 crc kubenswrapper[4967]: I1011 04:10:02.131690 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"470abdb5-fefb-48c8-bb4c-9ad0e81caedb","Type":"ContainerDied","Data":"5584621212aa51c0e1c6bdac7220d1190ccbbf6c7c75ee751ac70c2bce5cf42d"} Oct 11 04:10:02 crc kubenswrapper[4967]: I1011 04:10:02.131719 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"470abdb5-fefb-48c8-bb4c-9ad0e81caedb","Type":"ContainerDied","Data":"19de676933f0fbabbd00149485dff306db57a2781282138364b5f63f2e7a4a70"} Oct 11 04:10:02 crc kubenswrapper[4967]: I1011 04:10:02.134753 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="7fd48f8c-e380-44cb-8c5b-8380b637e0ef" containerName="nova-scheduler-scheduler" containerID="cri-o://4726e4a657dc964b9f92f99fa450dccebcb517bf23d903b7075ee3e034e8ccd1" gracePeriod=30 Oct 11 04:10:02 crc kubenswrapper[4967]: I1011 04:10:02.246131 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 11 04:10:02 crc kubenswrapper[4967]: I1011 04:10:02.256328 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ironic-inspector-0"] Oct 11 04:10:02 crc kubenswrapper[4967]: W1011 04:10:02.273745 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc619427d_62f7_463c_8262_d3b9fdd6ade0.slice/crio-80a08e28ce0d81e2b2219b6a984a6a2f0f0f339144348747b760ab9f53ac6129 WatchSource:0}: Error finding container 80a08e28ce0d81e2b2219b6a984a6a2f0f0f339144348747b760ab9f53ac6129: Status 404 returned error can't find the container with id 80a08e28ce0d81e2b2219b6a984a6a2f0f0f339144348747b760ab9f53ac6129 Oct 11 04:10:02 crc kubenswrapper[4967]: I1011 04:10:02.373699 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/470abdb5-fefb-48c8-bb4c-9ad0e81caedb-combined-ca-bundle\") pod \"470abdb5-fefb-48c8-bb4c-9ad0e81caedb\" (UID: \"470abdb5-fefb-48c8-bb4c-9ad0e81caedb\") " Oct 11 04:10:02 crc kubenswrapper[4967]: I1011 04:10:02.373907 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/470abdb5-fefb-48c8-bb4c-9ad0e81caedb-config-data\") pod \"470abdb5-fefb-48c8-bb4c-9ad0e81caedb\" (UID: \"470abdb5-fefb-48c8-bb4c-9ad0e81caedb\") " Oct 11 04:10:02 crc kubenswrapper[4967]: I1011 04:10:02.373961 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ccbbp\" (UniqueName: \"kubernetes.io/projected/470abdb5-fefb-48c8-bb4c-9ad0e81caedb-kube-api-access-ccbbp\") pod \"470abdb5-fefb-48c8-bb4c-9ad0e81caedb\" (UID: \"470abdb5-fefb-48c8-bb4c-9ad0e81caedb\") " Oct 11 04:10:02 crc kubenswrapper[4967]: I1011 04:10:02.374192 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/470abdb5-fefb-48c8-bb4c-9ad0e81caedb-logs\") pod \"470abdb5-fefb-48c8-bb4c-9ad0e81caedb\" (UID: \"470abdb5-fefb-48c8-bb4c-9ad0e81caedb\") " Oct 11 04:10:02 crc kubenswrapper[4967]: I1011 04:10:02.374471 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/470abdb5-fefb-48c8-bb4c-9ad0e81caedb-logs" (OuterVolumeSpecName: "logs") pod "470abdb5-fefb-48c8-bb4c-9ad0e81caedb" (UID: "470abdb5-fefb-48c8-bb4c-9ad0e81caedb"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:10:02 crc kubenswrapper[4967]: I1011 04:10:02.374958 4967 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/470abdb5-fefb-48c8-bb4c-9ad0e81caedb-logs\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:02 crc kubenswrapper[4967]: I1011 04:10:02.380313 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/470abdb5-fefb-48c8-bb4c-9ad0e81caedb-kube-api-access-ccbbp" (OuterVolumeSpecName: "kube-api-access-ccbbp") pod "470abdb5-fefb-48c8-bb4c-9ad0e81caedb" (UID: "470abdb5-fefb-48c8-bb4c-9ad0e81caedb"). InnerVolumeSpecName "kube-api-access-ccbbp". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:10:02 crc kubenswrapper[4967]: I1011 04:10:02.404706 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/470abdb5-fefb-48c8-bb4c-9ad0e81caedb-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "470abdb5-fefb-48c8-bb4c-9ad0e81caedb" (UID: "470abdb5-fefb-48c8-bb4c-9ad0e81caedb"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:02 crc kubenswrapper[4967]: I1011 04:10:02.406467 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/470abdb5-fefb-48c8-bb4c-9ad0e81caedb-config-data" (OuterVolumeSpecName: "config-data") pod "470abdb5-fefb-48c8-bb4c-9ad0e81caedb" (UID: "470abdb5-fefb-48c8-bb4c-9ad0e81caedb"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:02 crc kubenswrapper[4967]: I1011 04:10:02.476557 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/470abdb5-fefb-48c8-bb4c-9ad0e81caedb-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:02 crc kubenswrapper[4967]: I1011 04:10:02.476592 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/470abdb5-fefb-48c8-bb4c-9ad0e81caedb-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:02 crc kubenswrapper[4967]: I1011 04:10:02.476604 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ccbbp\" (UniqueName: \"kubernetes.io/projected/470abdb5-fefb-48c8-bb4c-9ad0e81caedb-kube-api-access-ccbbp\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:02 crc kubenswrapper[4967]: I1011 04:10:02.825241 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60a1ad86-dc2a-46a3-8c63-6e48d3a2d291" path="/var/lib/kubelet/pods/60a1ad86-dc2a-46a3-8c63-6e48d3a2d291/volumes" Oct 11 04:10:02 crc kubenswrapper[4967]: I1011 04:10:02.826059 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="95c0b2ec-d176-4832-897b-7c33176243cd" path="/var/lib/kubelet/pods/95c0b2ec-d176-4832-897b-7c33176243cd/volumes" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.144411 4967 generic.go:334] "Generic (PLEG): container finished" podID="c619427d-62f7-463c-8262-d3b9fdd6ade0" containerID="cbb0d40cc75ddfad366493c55e0980ae13aa99f41b64729d79547eb74905a28f" exitCode=0 Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.144518 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"c619427d-62f7-463c-8262-d3b9fdd6ade0","Type":"ContainerDied","Data":"cbb0d40cc75ddfad366493c55e0980ae13aa99f41b64729d79547eb74905a28f"} Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.144547 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"c619427d-62f7-463c-8262-d3b9fdd6ade0","Type":"ContainerStarted","Data":"80a08e28ce0d81e2b2219b6a984a6a2f0f0f339144348747b760ab9f53ac6129"} Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.150580 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"470abdb5-fefb-48c8-bb4c-9ad0e81caedb","Type":"ContainerDied","Data":"7a9539f736ba9147c2dffe84ce3e68562bedbcbbdddb76b85bc380fd37718743"} Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.150625 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.150628 4967 scope.go:117] "RemoveContainer" containerID="5584621212aa51c0e1c6bdac7220d1190ccbbf6c7c75ee751ac70c2bce5cf42d" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.150726 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="cee26db2-5786-43ed-8a26-3b0e55ca2964" containerName="nova-api-log" containerID="cri-o://409c9974b4f4831aab22106bce09e00932e188a272394ae0cf2cd9a2dcaa84f4" gracePeriod=30 Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.150848 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="cee26db2-5786-43ed-8a26-3b0e55ca2964" containerName="nova-api-api" containerID="cri-o://aba70ed5fba8bf5f924ca4861b46c22417dbe426f6ec5a6b879cb253cdece0ef" gracePeriod=30 Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.178489 4967 scope.go:117] "RemoveContainer" containerID="19de676933f0fbabbd00149485dff306db57a2781282138364b5f63f2e7a4a70" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.215133 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.225158 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.236692 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 11 04:10:03 crc kubenswrapper[4967]: E1011 04:10:03.237122 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="470abdb5-fefb-48c8-bb4c-9ad0e81caedb" containerName="nova-metadata-log" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.237140 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="470abdb5-fefb-48c8-bb4c-9ad0e81caedb" containerName="nova-metadata-log" Oct 11 04:10:03 crc kubenswrapper[4967]: E1011 04:10:03.237152 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="470abdb5-fefb-48c8-bb4c-9ad0e81caedb" containerName="nova-metadata-metadata" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.237159 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="470abdb5-fefb-48c8-bb4c-9ad0e81caedb" containerName="nova-metadata-metadata" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.237352 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="470abdb5-fefb-48c8-bb4c-9ad0e81caedb" containerName="nova-metadata-metadata" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.237387 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="470abdb5-fefb-48c8-bb4c-9ad0e81caedb" containerName="nova-metadata-log" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.238379 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.240533 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.243356 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.270417 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.398556 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebd723b4-b11e-46f6-a034-80269be7c26b-config-data\") pod \"nova-metadata-0\" (UID: \"ebd723b4-b11e-46f6-a034-80269be7c26b\") " pod="openstack/nova-metadata-0" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.398604 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rmfzr\" (UniqueName: \"kubernetes.io/projected/ebd723b4-b11e-46f6-a034-80269be7c26b-kube-api-access-rmfzr\") pod \"nova-metadata-0\" (UID: \"ebd723b4-b11e-46f6-a034-80269be7c26b\") " pod="openstack/nova-metadata-0" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.398684 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebd723b4-b11e-46f6-a034-80269be7c26b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ebd723b4-b11e-46f6-a034-80269be7c26b\") " pod="openstack/nova-metadata-0" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.398752 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebd723b4-b11e-46f6-a034-80269be7c26b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ebd723b4-b11e-46f6-a034-80269be7c26b\") " pod="openstack/nova-metadata-0" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.398809 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebd723b4-b11e-46f6-a034-80269be7c26b-logs\") pod \"nova-metadata-0\" (UID: \"ebd723b4-b11e-46f6-a034-80269be7c26b\") " pod="openstack/nova-metadata-0" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.501716 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebd723b4-b11e-46f6-a034-80269be7c26b-config-data\") pod \"nova-metadata-0\" (UID: \"ebd723b4-b11e-46f6-a034-80269be7c26b\") " pod="openstack/nova-metadata-0" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.501764 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rmfzr\" (UniqueName: \"kubernetes.io/projected/ebd723b4-b11e-46f6-a034-80269be7c26b-kube-api-access-rmfzr\") pod \"nova-metadata-0\" (UID: \"ebd723b4-b11e-46f6-a034-80269be7c26b\") " pod="openstack/nova-metadata-0" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.501829 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebd723b4-b11e-46f6-a034-80269be7c26b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ebd723b4-b11e-46f6-a034-80269be7c26b\") " pod="openstack/nova-metadata-0" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.501880 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebd723b4-b11e-46f6-a034-80269be7c26b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ebd723b4-b11e-46f6-a034-80269be7c26b\") " pod="openstack/nova-metadata-0" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.501926 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebd723b4-b11e-46f6-a034-80269be7c26b-logs\") pod \"nova-metadata-0\" (UID: \"ebd723b4-b11e-46f6-a034-80269be7c26b\") " pod="openstack/nova-metadata-0" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.502852 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebd723b4-b11e-46f6-a034-80269be7c26b-logs\") pod \"nova-metadata-0\" (UID: \"ebd723b4-b11e-46f6-a034-80269be7c26b\") " pod="openstack/nova-metadata-0" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.507511 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebd723b4-b11e-46f6-a034-80269be7c26b-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"ebd723b4-b11e-46f6-a034-80269be7c26b\") " pod="openstack/nova-metadata-0" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.507680 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebd723b4-b11e-46f6-a034-80269be7c26b-config-data\") pod \"nova-metadata-0\" (UID: \"ebd723b4-b11e-46f6-a034-80269be7c26b\") " pod="openstack/nova-metadata-0" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.507840 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebd723b4-b11e-46f6-a034-80269be7c26b-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"ebd723b4-b11e-46f6-a034-80269be7c26b\") " pod="openstack/nova-metadata-0" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.532553 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rmfzr\" (UniqueName: \"kubernetes.io/projected/ebd723b4-b11e-46f6-a034-80269be7c26b-kube-api-access-rmfzr\") pod \"nova-metadata-0\" (UID: \"ebd723b4-b11e-46f6-a034-80269be7c26b\") " pod="openstack/nova-metadata-0" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.607492 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.741434 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.815480 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-m98f2\" (UniqueName: \"kubernetes.io/projected/cee26db2-5786-43ed-8a26-3b0e55ca2964-kube-api-access-m98f2\") pod \"cee26db2-5786-43ed-8a26-3b0e55ca2964\" (UID: \"cee26db2-5786-43ed-8a26-3b0e55ca2964\") " Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.820694 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cee26db2-5786-43ed-8a26-3b0e55ca2964-kube-api-access-m98f2" (OuterVolumeSpecName: "kube-api-access-m98f2") pod "cee26db2-5786-43ed-8a26-3b0e55ca2964" (UID: "cee26db2-5786-43ed-8a26-3b0e55ca2964"). InnerVolumeSpecName "kube-api-access-m98f2". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.917781 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cee26db2-5786-43ed-8a26-3b0e55ca2964-logs\") pod \"cee26db2-5786-43ed-8a26-3b0e55ca2964\" (UID: \"cee26db2-5786-43ed-8a26-3b0e55ca2964\") " Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.917837 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cee26db2-5786-43ed-8a26-3b0e55ca2964-config-data\") pod \"cee26db2-5786-43ed-8a26-3b0e55ca2964\" (UID: \"cee26db2-5786-43ed-8a26-3b0e55ca2964\") " Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.917883 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cee26db2-5786-43ed-8a26-3b0e55ca2964-combined-ca-bundle\") pod \"cee26db2-5786-43ed-8a26-3b0e55ca2964\" (UID: \"cee26db2-5786-43ed-8a26-3b0e55ca2964\") " Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.919269 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/cee26db2-5786-43ed-8a26-3b0e55ca2964-logs" (OuterVolumeSpecName: "logs") pod "cee26db2-5786-43ed-8a26-3b0e55ca2964" (UID: "cee26db2-5786-43ed-8a26-3b0e55ca2964"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.919507 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-m98f2\" (UniqueName: \"kubernetes.io/projected/cee26db2-5786-43ed-8a26-3b0e55ca2964-kube-api-access-m98f2\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.947806 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cee26db2-5786-43ed-8a26-3b0e55ca2964-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "cee26db2-5786-43ed-8a26-3b0e55ca2964" (UID: "cee26db2-5786-43ed-8a26-3b0e55ca2964"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:03 crc kubenswrapper[4967]: I1011 04:10:03.948288 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cee26db2-5786-43ed-8a26-3b0e55ca2964-config-data" (OuterVolumeSpecName: "config-data") pod "cee26db2-5786-43ed-8a26-3b0e55ca2964" (UID: "cee26db2-5786-43ed-8a26-3b0e55ca2964"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.022171 4967 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/cee26db2-5786-43ed-8a26-3b0e55ca2964-logs\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.022212 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/cee26db2-5786-43ed-8a26-3b0e55ca2964-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.022229 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/cee26db2-5786-43ed-8a26-3b0e55ca2964-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:04 crc kubenswrapper[4967]: W1011 04:10:04.066268 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podebd723b4_b11e_46f6_a034_80269be7c26b.slice/crio-19bbed4ff049637f8c993f16fc2b92da04c401a6a7d7c2912ee5992a218a7d17 WatchSource:0}: Error finding container 19bbed4ff049637f8c993f16fc2b92da04c401a6a7d7c2912ee5992a218a7d17: Status 404 returned error can't find the container with id 19bbed4ff049637f8c993f16fc2b92da04c401a6a7d7c2912ee5992a218a7d17 Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.071426 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.101583 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.170019 4967 generic.go:334] "Generic (PLEG): container finished" podID="cee26db2-5786-43ed-8a26-3b0e55ca2964" containerID="aba70ed5fba8bf5f924ca4861b46c22417dbe426f6ec5a6b879cb253cdece0ef" exitCode=0 Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.170066 4967 generic.go:334] "Generic (PLEG): container finished" podID="cee26db2-5786-43ed-8a26-3b0e55ca2964" containerID="409c9974b4f4831aab22106bce09e00932e188a272394ae0cf2cd9a2dcaa84f4" exitCode=143 Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.170226 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cee26db2-5786-43ed-8a26-3b0e55ca2964","Type":"ContainerDied","Data":"aba70ed5fba8bf5f924ca4861b46c22417dbe426f6ec5a6b879cb253cdece0ef"} Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.170260 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cee26db2-5786-43ed-8a26-3b0e55ca2964","Type":"ContainerDied","Data":"409c9974b4f4831aab22106bce09e00932e188a272394ae0cf2cd9a2dcaa84f4"} Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.170274 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"cee26db2-5786-43ed-8a26-3b0e55ca2964","Type":"ContainerDied","Data":"e1e2cff23ec999a1f5978cc9669d532364f773752e4c1b9d3bb294a9796aa86f"} Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.170319 4967 scope.go:117] "RemoveContainer" containerID="aba70ed5fba8bf5f924ca4861b46c22417dbe426f6ec5a6b879cb253cdece0ef" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.170631 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.176444 4967 generic.go:334] "Generic (PLEG): container finished" podID="702714d0-d935-49ae-b19f-c0f80acd7758" containerID="8bdf9296be5a3e6cedf0cf960a89608cbc5c90a2175a407f8d332f1f82d80cc8" exitCode=0 Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.176555 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-dt5p8" event={"ID":"702714d0-d935-49ae-b19f-c0f80acd7758","Type":"ContainerDied","Data":"8bdf9296be5a3e6cedf0cf960a89608cbc5c90a2175a407f8d332f1f82d80cc8"} Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.177901 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ebd723b4-b11e-46f6-a034-80269be7c26b","Type":"ContainerStarted","Data":"19bbed4ff049637f8c993f16fc2b92da04c401a6a7d7c2912ee5992a218a7d17"} Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.205833 4967 scope.go:117] "RemoveContainer" containerID="409c9974b4f4831aab22106bce09e00932e188a272394ae0cf2cd9a2dcaa84f4" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.229241 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.243939 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.262206 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 11 04:10:04 crc kubenswrapper[4967]: E1011 04:10:04.263597 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cee26db2-5786-43ed-8a26-3b0e55ca2964" containerName="nova-api-log" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.263615 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="cee26db2-5786-43ed-8a26-3b0e55ca2964" containerName="nova-api-log" Oct 11 04:10:04 crc kubenswrapper[4967]: E1011 04:10:04.263641 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cee26db2-5786-43ed-8a26-3b0e55ca2964" containerName="nova-api-api" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.263696 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="cee26db2-5786-43ed-8a26-3b0e55ca2964" containerName="nova-api-api" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.263948 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="cee26db2-5786-43ed-8a26-3b0e55ca2964" containerName="nova-api-api" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.263978 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="cee26db2-5786-43ed-8a26-3b0e55ca2964" containerName="nova-api-log" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.264283 4967 scope.go:117] "RemoveContainer" containerID="aba70ed5fba8bf5f924ca4861b46c22417dbe426f6ec5a6b879cb253cdece0ef" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.265123 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 11 04:10:04 crc kubenswrapper[4967]: E1011 04:10:04.265231 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aba70ed5fba8bf5f924ca4861b46c22417dbe426f6ec5a6b879cb253cdece0ef\": container with ID starting with aba70ed5fba8bf5f924ca4861b46c22417dbe426f6ec5a6b879cb253cdece0ef not found: ID does not exist" containerID="aba70ed5fba8bf5f924ca4861b46c22417dbe426f6ec5a6b879cb253cdece0ef" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.265263 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aba70ed5fba8bf5f924ca4861b46c22417dbe426f6ec5a6b879cb253cdece0ef"} err="failed to get container status \"aba70ed5fba8bf5f924ca4861b46c22417dbe426f6ec5a6b879cb253cdece0ef\": rpc error: code = NotFound desc = could not find container \"aba70ed5fba8bf5f924ca4861b46c22417dbe426f6ec5a6b879cb253cdece0ef\": container with ID starting with aba70ed5fba8bf5f924ca4861b46c22417dbe426f6ec5a6b879cb253cdece0ef not found: ID does not exist" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.265289 4967 scope.go:117] "RemoveContainer" containerID="409c9974b4f4831aab22106bce09e00932e188a272394ae0cf2cd9a2dcaa84f4" Oct 11 04:10:04 crc kubenswrapper[4967]: E1011 04:10:04.266118 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"409c9974b4f4831aab22106bce09e00932e188a272394ae0cf2cd9a2dcaa84f4\": container with ID starting with 409c9974b4f4831aab22106bce09e00932e188a272394ae0cf2cd9a2dcaa84f4 not found: ID does not exist" containerID="409c9974b4f4831aab22106bce09e00932e188a272394ae0cf2cd9a2dcaa84f4" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.266194 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"409c9974b4f4831aab22106bce09e00932e188a272394ae0cf2cd9a2dcaa84f4"} err="failed to get container status \"409c9974b4f4831aab22106bce09e00932e188a272394ae0cf2cd9a2dcaa84f4\": rpc error: code = NotFound desc = could not find container \"409c9974b4f4831aab22106bce09e00932e188a272394ae0cf2cd9a2dcaa84f4\": container with ID starting with 409c9974b4f4831aab22106bce09e00932e188a272394ae0cf2cd9a2dcaa84f4 not found: ID does not exist" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.266227 4967 scope.go:117] "RemoveContainer" containerID="aba70ed5fba8bf5f924ca4861b46c22417dbe426f6ec5a6b879cb253cdece0ef" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.266989 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.267894 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.269605 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aba70ed5fba8bf5f924ca4861b46c22417dbe426f6ec5a6b879cb253cdece0ef"} err="failed to get container status \"aba70ed5fba8bf5f924ca4861b46c22417dbe426f6ec5a6b879cb253cdece0ef\": rpc error: code = NotFound desc = could not find container \"aba70ed5fba8bf5f924ca4861b46c22417dbe426f6ec5a6b879cb253cdece0ef\": container with ID starting with aba70ed5fba8bf5f924ca4861b46c22417dbe426f6ec5a6b879cb253cdece0ef not found: ID does not exist" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.269639 4967 scope.go:117] "RemoveContainer" containerID="409c9974b4f4831aab22106bce09e00932e188a272394ae0cf2cd9a2dcaa84f4" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.273777 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"409c9974b4f4831aab22106bce09e00932e188a272394ae0cf2cd9a2dcaa84f4"} err="failed to get container status \"409c9974b4f4831aab22106bce09e00932e188a272394ae0cf2cd9a2dcaa84f4\": rpc error: code = NotFound desc = could not find container \"409c9974b4f4831aab22106bce09e00932e188a272394ae0cf2cd9a2dcaa84f4\": container with ID starting with 409c9974b4f4831aab22106bce09e00932e188a272394ae0cf2cd9a2dcaa84f4 not found: ID does not exist" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.431121 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.441281 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83d7fc84-7338-4ad6-bf9b-901fb22a0f11-logs\") pod \"nova-api-0\" (UID: \"83d7fc84-7338-4ad6-bf9b-901fb22a0f11\") " pod="openstack/nova-api-0" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.441331 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83d7fc84-7338-4ad6-bf9b-901fb22a0f11-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"83d7fc84-7338-4ad6-bf9b-901fb22a0f11\") " pod="openstack/nova-api-0" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.441427 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tscxj\" (UniqueName: \"kubernetes.io/projected/83d7fc84-7338-4ad6-bf9b-901fb22a0f11-kube-api-access-tscxj\") pod \"nova-api-0\" (UID: \"83d7fc84-7338-4ad6-bf9b-901fb22a0f11\") " pod="openstack/nova-api-0" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.441566 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83d7fc84-7338-4ad6-bf9b-901fb22a0f11-config-data\") pod \"nova-api-0\" (UID: \"83d7fc84-7338-4ad6-bf9b-901fb22a0f11\") " pod="openstack/nova-api-0" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.543018 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tscxj\" (UniqueName: \"kubernetes.io/projected/83d7fc84-7338-4ad6-bf9b-901fb22a0f11-kube-api-access-tscxj\") pod \"nova-api-0\" (UID: \"83d7fc84-7338-4ad6-bf9b-901fb22a0f11\") " pod="openstack/nova-api-0" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.543143 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83d7fc84-7338-4ad6-bf9b-901fb22a0f11-config-data\") pod \"nova-api-0\" (UID: \"83d7fc84-7338-4ad6-bf9b-901fb22a0f11\") " pod="openstack/nova-api-0" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.543208 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83d7fc84-7338-4ad6-bf9b-901fb22a0f11-logs\") pod \"nova-api-0\" (UID: \"83d7fc84-7338-4ad6-bf9b-901fb22a0f11\") " pod="openstack/nova-api-0" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.543250 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83d7fc84-7338-4ad6-bf9b-901fb22a0f11-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"83d7fc84-7338-4ad6-bf9b-901fb22a0f11\") " pod="openstack/nova-api-0" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.543807 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83d7fc84-7338-4ad6-bf9b-901fb22a0f11-logs\") pod \"nova-api-0\" (UID: \"83d7fc84-7338-4ad6-bf9b-901fb22a0f11\") " pod="openstack/nova-api-0" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.550008 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83d7fc84-7338-4ad6-bf9b-901fb22a0f11-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"83d7fc84-7338-4ad6-bf9b-901fb22a0f11\") " pod="openstack/nova-api-0" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.550052 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83d7fc84-7338-4ad6-bf9b-901fb22a0f11-config-data\") pod \"nova-api-0\" (UID: \"83d7fc84-7338-4ad6-bf9b-901fb22a0f11\") " pod="openstack/nova-api-0" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.567987 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tscxj\" (UniqueName: \"kubernetes.io/projected/83d7fc84-7338-4ad6-bf9b-901fb22a0f11-kube-api-access-tscxj\") pod \"nova-api-0\" (UID: \"83d7fc84-7338-4ad6-bf9b-901fb22a0f11\") " pod="openstack/nova-api-0" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.583039 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.827885 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="470abdb5-fefb-48c8-bb4c-9ad0e81caedb" path="/var/lib/kubelet/pods/470abdb5-fefb-48c8-bb4c-9ad0e81caedb/volumes" Oct 11 04:10:04 crc kubenswrapper[4967]: I1011 04:10:04.829203 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cee26db2-5786-43ed-8a26-3b0e55ca2964" path="/var/lib/kubelet/pods/cee26db2-5786-43ed-8a26-3b0e55ca2964/volumes" Oct 11 04:10:05 crc kubenswrapper[4967]: I1011 04:10:05.094367 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 11 04:10:05 crc kubenswrapper[4967]: I1011 04:10:05.199486 4967 generic.go:334] "Generic (PLEG): container finished" podID="64c75ea6-aed6-4bf6-8eef-39e76384e14f" containerID="93d7b8ee4a62c1ecfe52437af2caf73381136995aee5a103e35e85022d1f814a" exitCode=0 Oct 11 04:10:05 crc kubenswrapper[4967]: I1011 04:10:05.199672 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"64c75ea6-aed6-4bf6-8eef-39e76384e14f","Type":"ContainerDied","Data":"93d7b8ee4a62c1ecfe52437af2caf73381136995aee5a103e35e85022d1f814a"} Oct 11 04:10:05 crc kubenswrapper[4967]: I1011 04:10:05.208634 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"83d7fc84-7338-4ad6-bf9b-901fb22a0f11","Type":"ContainerStarted","Data":"fc72043565edcb5f9f5a6bf9ed11a65f76939b465152281cd8555570f944dfbd"} Oct 11 04:10:05 crc kubenswrapper[4967]: I1011 04:10:05.217150 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ebd723b4-b11e-46f6-a034-80269be7c26b","Type":"ContainerStarted","Data":"eb9488bfd76965f1d7ccd7edd22ded7df84fe1b36de84d5b0693c3cae286ae2b"} Oct 11 04:10:05 crc kubenswrapper[4967]: I1011 04:10:05.217220 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ebd723b4-b11e-46f6-a034-80269be7c26b","Type":"ContainerStarted","Data":"ffe47789cdc470ff6ccfced73645ce45dc1e8c178037e40d42b7385ff255a77b"} Oct 11 04:10:05 crc kubenswrapper[4967]: I1011 04:10:05.265218 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.265197387 podStartE2EDuration="2.265197387s" podCreationTimestamp="2025-10-11 04:10:03 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:10:05.249703098 +0000 UTC m=+1133.212912031" watchObservedRunningTime="2025-10-11 04:10:05.265197387 +0000 UTC m=+1133.228406320" Oct 11 04:10:05 crc kubenswrapper[4967]: I1011 04:10:05.649825 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-dt5p8" Oct 11 04:10:05 crc kubenswrapper[4967]: I1011 04:10:05.763458 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/702714d0-d935-49ae-b19f-c0f80acd7758-combined-ca-bundle\") pod \"702714d0-d935-49ae-b19f-c0f80acd7758\" (UID: \"702714d0-d935-49ae-b19f-c0f80acd7758\") " Oct 11 04:10:05 crc kubenswrapper[4967]: I1011 04:10:05.763701 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/702714d0-d935-49ae-b19f-c0f80acd7758-scripts\") pod \"702714d0-d935-49ae-b19f-c0f80acd7758\" (UID: \"702714d0-d935-49ae-b19f-c0f80acd7758\") " Oct 11 04:10:05 crc kubenswrapper[4967]: I1011 04:10:05.763762 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-phn76\" (UniqueName: \"kubernetes.io/projected/702714d0-d935-49ae-b19f-c0f80acd7758-kube-api-access-phn76\") pod \"702714d0-d935-49ae-b19f-c0f80acd7758\" (UID: \"702714d0-d935-49ae-b19f-c0f80acd7758\") " Oct 11 04:10:05 crc kubenswrapper[4967]: I1011 04:10:05.763926 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/702714d0-d935-49ae-b19f-c0f80acd7758-config-data\") pod \"702714d0-d935-49ae-b19f-c0f80acd7758\" (UID: \"702714d0-d935-49ae-b19f-c0f80acd7758\") " Oct 11 04:10:05 crc kubenswrapper[4967]: I1011 04:10:05.768521 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/702714d0-d935-49ae-b19f-c0f80acd7758-kube-api-access-phn76" (OuterVolumeSpecName: "kube-api-access-phn76") pod "702714d0-d935-49ae-b19f-c0f80acd7758" (UID: "702714d0-d935-49ae-b19f-c0f80acd7758"). InnerVolumeSpecName "kube-api-access-phn76". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:10:05 crc kubenswrapper[4967]: I1011 04:10:05.769099 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/702714d0-d935-49ae-b19f-c0f80acd7758-scripts" (OuterVolumeSpecName: "scripts") pod "702714d0-d935-49ae-b19f-c0f80acd7758" (UID: "702714d0-d935-49ae-b19f-c0f80acd7758"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:05 crc kubenswrapper[4967]: I1011 04:10:05.795690 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/702714d0-d935-49ae-b19f-c0f80acd7758-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "702714d0-d935-49ae-b19f-c0f80acd7758" (UID: "702714d0-d935-49ae-b19f-c0f80acd7758"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:05 crc kubenswrapper[4967]: I1011 04:10:05.803332 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/702714d0-d935-49ae-b19f-c0f80acd7758-config-data" (OuterVolumeSpecName: "config-data") pod "702714d0-d935-49ae-b19f-c0f80acd7758" (UID: "702714d0-d935-49ae-b19f-c0f80acd7758"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:05 crc kubenswrapper[4967]: I1011 04:10:05.872787 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/702714d0-d935-49ae-b19f-c0f80acd7758-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:05 crc kubenswrapper[4967]: I1011 04:10:05.873188 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-phn76\" (UniqueName: \"kubernetes.io/projected/702714d0-d935-49ae-b19f-c0f80acd7758-kube-api-access-phn76\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:05 crc kubenswrapper[4967]: I1011 04:10:05.873302 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/702714d0-d935-49ae-b19f-c0f80acd7758-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:05 crc kubenswrapper[4967]: I1011 04:10:05.873472 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/702714d0-d935-49ae-b19f-c0f80acd7758-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:06 crc kubenswrapper[4967]: I1011 04:10:06.267117 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-db-sync-dt5p8" event={"ID":"702714d0-d935-49ae-b19f-c0f80acd7758","Type":"ContainerDied","Data":"7e6407148046234336abdd1c59ec23032385988bf042f14b941198b1e42c9317"} Oct 11 04:10:06 crc kubenswrapper[4967]: I1011 04:10:06.267164 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="7e6407148046234336abdd1c59ec23032385988bf042f14b941198b1e42c9317" Oct 11 04:10:06 crc kubenswrapper[4967]: I1011 04:10:06.267245 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-db-sync-dt5p8" Oct 11 04:10:06 crc kubenswrapper[4967]: I1011 04:10:06.272689 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"83d7fc84-7338-4ad6-bf9b-901fb22a0f11","Type":"ContainerStarted","Data":"d82b5d8986efc77f8f88dc4bda177c119d00d1824a5b9b80a3dbdff0cbcbdde9"} Oct 11 04:10:06 crc kubenswrapper[4967]: I1011 04:10:06.315193 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 11 04:10:06 crc kubenswrapper[4967]: E1011 04:10:06.315697 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="702714d0-d935-49ae-b19f-c0f80acd7758" containerName="nova-cell1-conductor-db-sync" Oct 11 04:10:06 crc kubenswrapper[4967]: I1011 04:10:06.315721 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="702714d0-d935-49ae-b19f-c0f80acd7758" containerName="nova-cell1-conductor-db-sync" Oct 11 04:10:06 crc kubenswrapper[4967]: I1011 04:10:06.315967 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="702714d0-d935-49ae-b19f-c0f80acd7758" containerName="nova-cell1-conductor-db-sync" Oct 11 04:10:06 crc kubenswrapper[4967]: I1011 04:10:06.316904 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 11 04:10:06 crc kubenswrapper[4967]: I1011 04:10:06.327985 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-conductor-config-data" Oct 11 04:10:06 crc kubenswrapper[4967]: I1011 04:10:06.332327 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 11 04:10:06 crc kubenswrapper[4967]: I1011 04:10:06.388785 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9b2e413-f97a-4b67-9071-d1aed0db806f-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"a9b2e413-f97a-4b67-9071-d1aed0db806f\") " pod="openstack/nova-cell1-conductor-0" Oct 11 04:10:06 crc kubenswrapper[4967]: I1011 04:10:06.388920 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2tw6v\" (UniqueName: \"kubernetes.io/projected/a9b2e413-f97a-4b67-9071-d1aed0db806f-kube-api-access-2tw6v\") pod \"nova-cell1-conductor-0\" (UID: \"a9b2e413-f97a-4b67-9071-d1aed0db806f\") " pod="openstack/nova-cell1-conductor-0" Oct 11 04:10:06 crc kubenswrapper[4967]: I1011 04:10:06.388978 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9b2e413-f97a-4b67-9071-d1aed0db806f-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"a9b2e413-f97a-4b67-9071-d1aed0db806f\") " pod="openstack/nova-cell1-conductor-0" Oct 11 04:10:06 crc kubenswrapper[4967]: I1011 04:10:06.490738 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9b2e413-f97a-4b67-9071-d1aed0db806f-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"a9b2e413-f97a-4b67-9071-d1aed0db806f\") " pod="openstack/nova-cell1-conductor-0" Oct 11 04:10:06 crc kubenswrapper[4967]: I1011 04:10:06.490861 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2tw6v\" (UniqueName: \"kubernetes.io/projected/a9b2e413-f97a-4b67-9071-d1aed0db806f-kube-api-access-2tw6v\") pod \"nova-cell1-conductor-0\" (UID: \"a9b2e413-f97a-4b67-9071-d1aed0db806f\") " pod="openstack/nova-cell1-conductor-0" Oct 11 04:10:06 crc kubenswrapper[4967]: I1011 04:10:06.490943 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9b2e413-f97a-4b67-9071-d1aed0db806f-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"a9b2e413-f97a-4b67-9071-d1aed0db806f\") " pod="openstack/nova-cell1-conductor-0" Oct 11 04:10:06 crc kubenswrapper[4967]: I1011 04:10:06.495173 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/a9b2e413-f97a-4b67-9071-d1aed0db806f-config-data\") pod \"nova-cell1-conductor-0\" (UID: \"a9b2e413-f97a-4b67-9071-d1aed0db806f\") " pod="openstack/nova-cell1-conductor-0" Oct 11 04:10:06 crc kubenswrapper[4967]: I1011 04:10:06.496804 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/a9b2e413-f97a-4b67-9071-d1aed0db806f-combined-ca-bundle\") pod \"nova-cell1-conductor-0\" (UID: \"a9b2e413-f97a-4b67-9071-d1aed0db806f\") " pod="openstack/nova-cell1-conductor-0" Oct 11 04:10:06 crc kubenswrapper[4967]: I1011 04:10:06.506030 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2tw6v\" (UniqueName: \"kubernetes.io/projected/a9b2e413-f97a-4b67-9071-d1aed0db806f-kube-api-access-2tw6v\") pod \"nova-cell1-conductor-0\" (UID: \"a9b2e413-f97a-4b67-9071-d1aed0db806f\") " pod="openstack/nova-cell1-conductor-0" Oct 11 04:10:06 crc kubenswrapper[4967]: I1011 04:10:06.648464 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-conductor-0" Oct 11 04:10:07 crc kubenswrapper[4967]: I1011 04:10:07.122001 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-conductor-0"] Oct 11 04:10:07 crc kubenswrapper[4967]: W1011 04:10:07.130317 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poda9b2e413_f97a_4b67_9071_d1aed0db806f.slice/crio-d1a2d6b5463ab9a1f7d6c798838459a4baab1425f70709335c8cb70959763403 WatchSource:0}: Error finding container d1a2d6b5463ab9a1f7d6c798838459a4baab1425f70709335c8cb70959763403: Status 404 returned error can't find the container with id d1a2d6b5463ab9a1f7d6c798838459a4baab1425f70709335c8cb70959763403 Oct 11 04:10:07 crc kubenswrapper[4967]: I1011 04:10:07.283179 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"a9b2e413-f97a-4b67-9071-d1aed0db806f","Type":"ContainerStarted","Data":"d1a2d6b5463ab9a1f7d6c798838459a4baab1425f70709335c8cb70959763403"} Oct 11 04:10:07 crc kubenswrapper[4967]: I1011 04:10:07.285384 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"83d7fc84-7338-4ad6-bf9b-901fb22a0f11","Type":"ContainerStarted","Data":"fce5a60c173cabbdb2679451ca90fa09a906e0f73b3616f8f59d78574a7b70d6"} Oct 11 04:10:07 crc kubenswrapper[4967]: I1011 04:10:07.313685 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.313668685 podStartE2EDuration="3.313668685s" podCreationTimestamp="2025-10-11 04:10:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:10:07.300463258 +0000 UTC m=+1135.263672191" watchObservedRunningTime="2025-10-11 04:10:07.313668685 +0000 UTC m=+1135.276877618" Oct 11 04:10:08 crc kubenswrapper[4967]: I1011 04:10:08.298141 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-conductor-0" event={"ID":"a9b2e413-f97a-4b67-9071-d1aed0db806f","Type":"ContainerStarted","Data":"f8c72f6d8369628256b5a74ab836faea017d026b0e244c318c5db6b9bb90f501"} Oct 11 04:10:08 crc kubenswrapper[4967]: I1011 04:10:08.608231 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 11 04:10:08 crc kubenswrapper[4967]: I1011 04:10:08.608336 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 11 04:10:09 crc kubenswrapper[4967]: I1011 04:10:09.336543 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-conductor-0" podStartSLOduration=3.33652455 podStartE2EDuration="3.33652455s" podCreationTimestamp="2025-10-11 04:10:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:10:09.335398803 +0000 UTC m=+1137.298607736" watchObservedRunningTime="2025-10-11 04:10:09.33652455 +0000 UTC m=+1137.299733493" Oct 11 04:10:11 crc kubenswrapper[4967]: I1011 04:10:11.648857 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-conductor-0" Oct 11 04:10:13 crc kubenswrapper[4967]: I1011 04:10:13.349745 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"64c75ea6-aed6-4bf6-8eef-39e76384e14f","Type":"ContainerStarted","Data":"ad668f94310d78ecf52994e730277185b48ab6fae339ec0352d134f3cf5dc2de"} Oct 11 04:10:13 crc kubenswrapper[4967]: I1011 04:10:13.352926 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"c619427d-62f7-463c-8262-d3b9fdd6ade0","Type":"ContainerStarted","Data":"f4ba2fa5da86ebf42130d9fbfd3ce21b85347b361dddd861ebc9c32d7af6d96c"} Oct 11 04:10:13 crc kubenswrapper[4967]: I1011 04:10:13.608645 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 11 04:10:13 crc kubenswrapper[4967]: I1011 04:10:13.608697 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 11 04:10:14 crc kubenswrapper[4967]: I1011 04:10:14.089354 4967 patch_prober.go:28] interesting pod/machine-config-daemon-sgrv2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:10:14 crc kubenswrapper[4967]: I1011 04:10:14.089674 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:10:14 crc kubenswrapper[4967]: I1011 04:10:14.089726 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" Oct 11 04:10:14 crc kubenswrapper[4967]: I1011 04:10:14.090559 4967 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"37a6f451ae95b67f0b65e3cc023ede1c1770c6a148d14dd88cd2fe303a4c8469"} pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 11 04:10:14 crc kubenswrapper[4967]: I1011 04:10:14.090629 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" containerID="cri-o://37a6f451ae95b67f0b65e3cc023ede1c1770c6a148d14dd88cd2fe303a4c8469" gracePeriod=600 Oct 11 04:10:14 crc kubenswrapper[4967]: I1011 04:10:14.363725 4967 generic.go:334] "Generic (PLEG): container finished" podID="c619427d-62f7-463c-8262-d3b9fdd6ade0" containerID="f4ba2fa5da86ebf42130d9fbfd3ce21b85347b361dddd861ebc9c32d7af6d96c" exitCode=0 Oct 11 04:10:14 crc kubenswrapper[4967]: I1011 04:10:14.363769 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"c619427d-62f7-463c-8262-d3b9fdd6ade0","Type":"ContainerDied","Data":"f4ba2fa5da86ebf42130d9fbfd3ce21b85347b361dddd861ebc9c32d7af6d96c"} Oct 11 04:10:14 crc kubenswrapper[4967]: I1011 04:10:14.584117 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 11 04:10:14 crc kubenswrapper[4967]: I1011 04:10:14.584176 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 11 04:10:14 crc kubenswrapper[4967]: I1011 04:10:14.629444 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ebd723b4-b11e-46f6-a034-80269be7c26b" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 11 04:10:14 crc kubenswrapper[4967]: I1011 04:10:14.629901 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="ebd723b4-b11e-46f6-a034-80269be7c26b" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 11 04:10:14 crc kubenswrapper[4967]: I1011 04:10:14.986872 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 11 04:10:15 crc kubenswrapper[4967]: I1011 04:10:15.374436 4967 generic.go:334] "Generic (PLEG): container finished" podID="97c07678-14be-410c-b61f-498cb49bc960" containerID="37a6f451ae95b67f0b65e3cc023ede1c1770c6a148d14dd88cd2fe303a4c8469" exitCode=0 Oct 11 04:10:15 crc kubenswrapper[4967]: I1011 04:10:15.374583 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" event={"ID":"97c07678-14be-410c-b61f-498cb49bc960","Type":"ContainerDied","Data":"37a6f451ae95b67f0b65e3cc023ede1c1770c6a148d14dd88cd2fe303a4c8469"} Oct 11 04:10:15 crc kubenswrapper[4967]: I1011 04:10:15.374798 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" event={"ID":"97c07678-14be-410c-b61f-498cb49bc960","Type":"ContainerStarted","Data":"c14dde3bbb545266e69254f79fceffdb0e7aa40ff2dad2c747eae9612aaecda9"} Oct 11 04:10:15 crc kubenswrapper[4967]: I1011 04:10:15.374827 4967 scope.go:117] "RemoveContainer" containerID="f1443b341ff7c816afffce3e40de74a6868efcc5c30c2eb7be83fc5ffc5860c7" Oct 11 04:10:15 crc kubenswrapper[4967]: I1011 04:10:15.380768 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"c619427d-62f7-463c-8262-d3b9fdd6ade0","Type":"ContainerStarted","Data":"b1d3e718d39b2cd1f8952fbdeb794fba3b6312bb9cc84c5f7a9f2c7f43c39058"} Oct 11 04:10:15 crc kubenswrapper[4967]: I1011 04:10:15.666298 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="83d7fc84-7338-4ad6-bf9b-901fb22a0f11" containerName="nova-api-api" probeResult="failure" output="Get \"http://10.217.0.203:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 11 04:10:15 crc kubenswrapper[4967]: I1011 04:10:15.666337 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="83d7fc84-7338-4ad6-bf9b-901fb22a0f11" containerName="nova-api-log" probeResult="failure" output="Get \"http://10.217.0.203:8774/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 11 04:10:16 crc kubenswrapper[4967]: I1011 04:10:16.392851 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"c619427d-62f7-463c-8262-d3b9fdd6ade0","Type":"ContainerStarted","Data":"3d9d1472710a3f430b843fb1e3c033c16c4461c5301408321b031dc3705a0a90"} Oct 11 04:10:16 crc kubenswrapper[4967]: I1011 04:10:16.687057 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-conductor-0" Oct 11 04:10:17 crc kubenswrapper[4967]: I1011 04:10:17.425660 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"c619427d-62f7-463c-8262-d3b9fdd6ade0","Type":"ContainerStarted","Data":"c66c96afc4454d2254f40542f7ec5041d3a43e303460b8d8a26b7c2e732bc529"} Oct 11 04:10:17 crc kubenswrapper[4967]: I1011 04:10:17.425700 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-inspector-0" event={"ID":"c619427d-62f7-463c-8262-d3b9fdd6ade0","Type":"ContainerStarted","Data":"1ebc95b42a3194d118a2c9c81d42c03901a34b476461f44eeeee42c2bc020f4d"} Oct 11 04:10:17 crc kubenswrapper[4967]: I1011 04:10:17.425798 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-inspector-0" Oct 11 04:10:17 crc kubenswrapper[4967]: I1011 04:10:17.469941 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-inspector-0" podStartSLOduration=7.018674801 podStartE2EDuration="16.46992225s" podCreationTimestamp="2025-10-11 04:10:01 +0000 UTC" firstStartedPulling="2025-10-11 04:10:03.147182448 +0000 UTC m=+1131.110391381" lastFinishedPulling="2025-10-11 04:10:12.598429887 +0000 UTC m=+1140.561638830" observedRunningTime="2025-10-11 04:10:17.463360618 +0000 UTC m=+1145.426569551" watchObservedRunningTime="2025-10-11 04:10:17.46992225 +0000 UTC m=+1145.433131183" Oct 11 04:10:21 crc kubenswrapper[4967]: I1011 04:10:21.724347 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-inspector-0" Oct 11 04:10:21 crc kubenswrapper[4967]: I1011 04:10:21.724727 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ironic-inspector-0" Oct 11 04:10:21 crc kubenswrapper[4967]: I1011 04:10:21.724742 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/ironic-inspector-0" Oct 11 04:10:21 crc kubenswrapper[4967]: I1011 04:10:21.724754 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-inspector-0" Oct 11 04:10:21 crc kubenswrapper[4967]: I1011 04:10:21.727605 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-inspector-0" Oct 11 04:10:21 crc kubenswrapper[4967]: I1011 04:10:21.756205 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ironic-inspector-0" Oct 11 04:10:21 crc kubenswrapper[4967]: I1011 04:10:21.758151 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/ironic-inspector-0" Oct 11 04:10:22 crc kubenswrapper[4967]: I1011 04:10:22.479523 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-inspector-0" Oct 11 04:10:22 crc kubenswrapper[4967]: I1011 04:10:22.481472 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-inspector-0" Oct 11 04:10:23 crc kubenswrapper[4967]: I1011 04:10:23.614044 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 11 04:10:23 crc kubenswrapper[4967]: I1011 04:10:23.615422 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 11 04:10:23 crc kubenswrapper[4967]: I1011 04:10:23.619815 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 11 04:10:24 crc kubenswrapper[4967]: I1011 04:10:24.496612 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 11 04:10:24 crc kubenswrapper[4967]: I1011 04:10:24.587831 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 11 04:10:24 crc kubenswrapper[4967]: I1011 04:10:24.589207 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 11 04:10:24 crc kubenswrapper[4967]: I1011 04:10:24.591555 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 11 04:10:24 crc kubenswrapper[4967]: I1011 04:10:24.615356 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 11 04:10:25 crc kubenswrapper[4967]: I1011 04:10:25.496666 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 11 04:10:25 crc kubenswrapper[4967]: I1011 04:10:25.501320 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 11 04:10:25 crc kubenswrapper[4967]: I1011 04:10:25.658006 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-87xhd"] Oct 11 04:10:25 crc kubenswrapper[4967]: I1011 04:10:25.662387 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-87xhd" Oct 11 04:10:25 crc kubenswrapper[4967]: I1011 04:10:25.700137 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-87xhd"] Oct 11 04:10:25 crc kubenswrapper[4967]: I1011 04:10:25.700850 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef02fdf0-4696-42b9-9095-b000a70fe334-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-87xhd\" (UID: \"ef02fdf0-4696-42b9-9095-b000a70fe334\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-87xhd" Oct 11 04:10:25 crc kubenswrapper[4967]: I1011 04:10:25.700991 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef02fdf0-4696-42b9-9095-b000a70fe334-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-87xhd\" (UID: \"ef02fdf0-4696-42b9-9095-b000a70fe334\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-87xhd" Oct 11 04:10:25 crc kubenswrapper[4967]: I1011 04:10:25.701098 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pkgj6\" (UniqueName: \"kubernetes.io/projected/ef02fdf0-4696-42b9-9095-b000a70fe334-kube-api-access-pkgj6\") pod \"dnsmasq-dns-cd5cbd7b9-87xhd\" (UID: \"ef02fdf0-4696-42b9-9095-b000a70fe334\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-87xhd" Oct 11 04:10:25 crc kubenswrapper[4967]: I1011 04:10:25.701218 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ef02fdf0-4696-42b9-9095-b000a70fe334-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-87xhd\" (UID: \"ef02fdf0-4696-42b9-9095-b000a70fe334\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-87xhd" Oct 11 04:10:25 crc kubenswrapper[4967]: I1011 04:10:25.701313 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef02fdf0-4696-42b9-9095-b000a70fe334-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-87xhd\" (UID: \"ef02fdf0-4696-42b9-9095-b000a70fe334\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-87xhd" Oct 11 04:10:25 crc kubenswrapper[4967]: I1011 04:10:25.701443 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef02fdf0-4696-42b9-9095-b000a70fe334-config\") pod \"dnsmasq-dns-cd5cbd7b9-87xhd\" (UID: \"ef02fdf0-4696-42b9-9095-b000a70fe334\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-87xhd" Oct 11 04:10:25 crc kubenswrapper[4967]: I1011 04:10:25.802628 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef02fdf0-4696-42b9-9095-b000a70fe334-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-87xhd\" (UID: \"ef02fdf0-4696-42b9-9095-b000a70fe334\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-87xhd" Oct 11 04:10:25 crc kubenswrapper[4967]: I1011 04:10:25.802908 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef02fdf0-4696-42b9-9095-b000a70fe334-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-87xhd\" (UID: \"ef02fdf0-4696-42b9-9095-b000a70fe334\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-87xhd" Oct 11 04:10:25 crc kubenswrapper[4967]: I1011 04:10:25.802936 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pkgj6\" (UniqueName: \"kubernetes.io/projected/ef02fdf0-4696-42b9-9095-b000a70fe334-kube-api-access-pkgj6\") pod \"dnsmasq-dns-cd5cbd7b9-87xhd\" (UID: \"ef02fdf0-4696-42b9-9095-b000a70fe334\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-87xhd" Oct 11 04:10:25 crc kubenswrapper[4967]: I1011 04:10:25.802973 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ef02fdf0-4696-42b9-9095-b000a70fe334-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-87xhd\" (UID: \"ef02fdf0-4696-42b9-9095-b000a70fe334\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-87xhd" Oct 11 04:10:25 crc kubenswrapper[4967]: I1011 04:10:25.802997 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef02fdf0-4696-42b9-9095-b000a70fe334-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-87xhd\" (UID: \"ef02fdf0-4696-42b9-9095-b000a70fe334\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-87xhd" Oct 11 04:10:25 crc kubenswrapper[4967]: I1011 04:10:25.803032 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef02fdf0-4696-42b9-9095-b000a70fe334-config\") pod \"dnsmasq-dns-cd5cbd7b9-87xhd\" (UID: \"ef02fdf0-4696-42b9-9095-b000a70fe334\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-87xhd" Oct 11 04:10:25 crc kubenswrapper[4967]: I1011 04:10:25.804117 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/ef02fdf0-4696-42b9-9095-b000a70fe334-config\") pod \"dnsmasq-dns-cd5cbd7b9-87xhd\" (UID: \"ef02fdf0-4696-42b9-9095-b000a70fe334\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-87xhd" Oct 11 04:10:25 crc kubenswrapper[4967]: I1011 04:10:25.804383 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/ef02fdf0-4696-42b9-9095-b000a70fe334-dns-svc\") pod \"dnsmasq-dns-cd5cbd7b9-87xhd\" (UID: \"ef02fdf0-4696-42b9-9095-b000a70fe334\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-87xhd" Oct 11 04:10:25 crc kubenswrapper[4967]: I1011 04:10:25.804671 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/ef02fdf0-4696-42b9-9095-b000a70fe334-ovsdbserver-sb\") pod \"dnsmasq-dns-cd5cbd7b9-87xhd\" (UID: \"ef02fdf0-4696-42b9-9095-b000a70fe334\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-87xhd" Oct 11 04:10:25 crc kubenswrapper[4967]: I1011 04:10:25.805365 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/ef02fdf0-4696-42b9-9095-b000a70fe334-dns-swift-storage-0\") pod \"dnsmasq-dns-cd5cbd7b9-87xhd\" (UID: \"ef02fdf0-4696-42b9-9095-b000a70fe334\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-87xhd" Oct 11 04:10:25 crc kubenswrapper[4967]: I1011 04:10:25.805376 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/ef02fdf0-4696-42b9-9095-b000a70fe334-ovsdbserver-nb\") pod \"dnsmasq-dns-cd5cbd7b9-87xhd\" (UID: \"ef02fdf0-4696-42b9-9095-b000a70fe334\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-87xhd" Oct 11 04:10:25 crc kubenswrapper[4967]: I1011 04:10:25.823111 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pkgj6\" (UniqueName: \"kubernetes.io/projected/ef02fdf0-4696-42b9-9095-b000a70fe334-kube-api-access-pkgj6\") pod \"dnsmasq-dns-cd5cbd7b9-87xhd\" (UID: \"ef02fdf0-4696-42b9-9095-b000a70fe334\") " pod="openstack/dnsmasq-dns-cd5cbd7b9-87xhd" Oct 11 04:10:25 crc kubenswrapper[4967]: I1011 04:10:25.988586 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-cd5cbd7b9-87xhd" Oct 11 04:10:26 crc kubenswrapper[4967]: I1011 04:10:26.503283 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/dnsmasq-dns-cd5cbd7b9-87xhd"] Oct 11 04:10:27 crc kubenswrapper[4967]: I1011 04:10:27.517551 4967 generic.go:334] "Generic (PLEG): container finished" podID="ef02fdf0-4696-42b9-9095-b000a70fe334" containerID="8e88f9fb1aeb3011e7df85c9fa63cd2c3eb5313937071c3d300caefd1fae0404" exitCode=0 Oct 11 04:10:27 crc kubenswrapper[4967]: I1011 04:10:27.517699 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-87xhd" event={"ID":"ef02fdf0-4696-42b9-9095-b000a70fe334","Type":"ContainerDied","Data":"8e88f9fb1aeb3011e7df85c9fa63cd2c3eb5313937071c3d300caefd1fae0404"} Oct 11 04:10:27 crc kubenswrapper[4967]: I1011 04:10:27.518306 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-87xhd" event={"ID":"ef02fdf0-4696-42b9-9095-b000a70fe334","Type":"ContainerStarted","Data":"da58f23ca4a1283da6d8654c5648bf40025305f1d0022c462302eb05f361f548"} Oct 11 04:10:27 crc kubenswrapper[4967]: I1011 04:10:27.656784 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:10:27 crc kubenswrapper[4967]: I1011 04:10:27.657113 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3e95b396-4e67-4df5-a39a-15ff3a00e6a1" containerName="ceilometer-central-agent" containerID="cri-o://f7362d856d4a38869ca9e64c37955f1a0544ca77568782afb13d718ba74b351a" gracePeriod=30 Oct 11 04:10:27 crc kubenswrapper[4967]: I1011 04:10:27.657632 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3e95b396-4e67-4df5-a39a-15ff3a00e6a1" containerName="proxy-httpd" containerID="cri-o://7da08d64f43a6586cbd116ace09f566fd2c959ac5ecddb850663fc0be41dc2f2" gracePeriod=30 Oct 11 04:10:27 crc kubenswrapper[4967]: I1011 04:10:27.657711 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3e95b396-4e67-4df5-a39a-15ff3a00e6a1" containerName="ceilometer-notification-agent" containerID="cri-o://0875cf7cbc65df1e3e6e99f37987fe6f36adf3677c200154187ee9d1072fac73" gracePeriod=30 Oct 11 04:10:27 crc kubenswrapper[4967]: I1011 04:10:27.657790 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/ceilometer-0" podUID="3e95b396-4e67-4df5-a39a-15ff3a00e6a1" containerName="sg-core" containerID="cri-o://d8cd6304732e32865c9d7f503d2a7316b66b5d1512696297666da84a07823c77" gracePeriod=30 Oct 11 04:10:28 crc kubenswrapper[4967]: I1011 04:10:28.026004 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 11 04:10:28 crc kubenswrapper[4967]: I1011 04:10:28.529348 4967 generic.go:334] "Generic (PLEG): container finished" podID="3e95b396-4e67-4df5-a39a-15ff3a00e6a1" containerID="7da08d64f43a6586cbd116ace09f566fd2c959ac5ecddb850663fc0be41dc2f2" exitCode=0 Oct 11 04:10:28 crc kubenswrapper[4967]: I1011 04:10:28.529387 4967 generic.go:334] "Generic (PLEG): container finished" podID="3e95b396-4e67-4df5-a39a-15ff3a00e6a1" containerID="d8cd6304732e32865c9d7f503d2a7316b66b5d1512696297666da84a07823c77" exitCode=2 Oct 11 04:10:28 crc kubenswrapper[4967]: I1011 04:10:28.529399 4967 generic.go:334] "Generic (PLEG): container finished" podID="3e95b396-4e67-4df5-a39a-15ff3a00e6a1" containerID="f7362d856d4a38869ca9e64c37955f1a0544ca77568782afb13d718ba74b351a" exitCode=0 Oct 11 04:10:28 crc kubenswrapper[4967]: I1011 04:10:28.529438 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e95b396-4e67-4df5-a39a-15ff3a00e6a1","Type":"ContainerDied","Data":"7da08d64f43a6586cbd116ace09f566fd2c959ac5ecddb850663fc0be41dc2f2"} Oct 11 04:10:28 crc kubenswrapper[4967]: I1011 04:10:28.529466 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e95b396-4e67-4df5-a39a-15ff3a00e6a1","Type":"ContainerDied","Data":"d8cd6304732e32865c9d7f503d2a7316b66b5d1512696297666da84a07823c77"} Oct 11 04:10:28 crc kubenswrapper[4967]: I1011 04:10:28.529477 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e95b396-4e67-4df5-a39a-15ff3a00e6a1","Type":"ContainerDied","Data":"f7362d856d4a38869ca9e64c37955f1a0544ca77568782afb13d718ba74b351a"} Oct 11 04:10:28 crc kubenswrapper[4967]: I1011 04:10:28.534272 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="83d7fc84-7338-4ad6-bf9b-901fb22a0f11" containerName="nova-api-log" containerID="cri-o://d82b5d8986efc77f8f88dc4bda177c119d00d1824a5b9b80a3dbdff0cbcbdde9" gracePeriod=30 Oct 11 04:10:28 crc kubenswrapper[4967]: I1011 04:10:28.535093 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-cd5cbd7b9-87xhd" event={"ID":"ef02fdf0-4696-42b9-9095-b000a70fe334","Type":"ContainerStarted","Data":"822ed4211cfb294664df282e1dcae12ef5e0d82aa8d235a03115847d2c15710e"} Oct 11 04:10:28 crc kubenswrapper[4967]: I1011 04:10:28.535132 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/dnsmasq-dns-cd5cbd7b9-87xhd" Oct 11 04:10:28 crc kubenswrapper[4967]: I1011 04:10:28.535450 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="83d7fc84-7338-4ad6-bf9b-901fb22a0f11" containerName="nova-api-api" containerID="cri-o://fce5a60c173cabbdb2679451ca90fa09a906e0f73b3616f8f59d78574a7b70d6" gracePeriod=30 Oct 11 04:10:28 crc kubenswrapper[4967]: I1011 04:10:28.580089 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/dnsmasq-dns-cd5cbd7b9-87xhd" podStartSLOduration=3.580048185 podStartE2EDuration="3.580048185s" podCreationTimestamp="2025-10-11 04:10:25 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:10:28.553811677 +0000 UTC m=+1156.517020600" watchObservedRunningTime="2025-10-11 04:10:28.580048185 +0000 UTC m=+1156.543257118" Oct 11 04:10:29 crc kubenswrapper[4967]: I1011 04:10:29.546100 4967 generic.go:334] "Generic (PLEG): container finished" podID="83d7fc84-7338-4ad6-bf9b-901fb22a0f11" containerID="d82b5d8986efc77f8f88dc4bda177c119d00d1824a5b9b80a3dbdff0cbcbdde9" exitCode=143 Oct 11 04:10:29 crc kubenswrapper[4967]: I1011 04:10:29.546297 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"83d7fc84-7338-4ad6-bf9b-901fb22a0f11","Type":"ContainerDied","Data":"d82b5d8986efc77f8f88dc4bda177c119d00d1824a5b9b80a3dbdff0cbcbdde9"} Oct 11 04:10:30 crc kubenswrapper[4967]: I1011 04:10:30.556689 4967 generic.go:334] "Generic (PLEG): container finished" podID="1317501c-8db9-4f03-a1cf-10e92b2ccc8b" containerID="fb17607f6184c47fb87463585b50b5bac67a191f9c083416e9f65b664822e600" exitCode=137 Oct 11 04:10:30 crc kubenswrapper[4967]: I1011 04:10:30.556762 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1317501c-8db9-4f03-a1cf-10e92b2ccc8b","Type":"ContainerDied","Data":"fb17607f6184c47fb87463585b50b5bac67a191f9c083416e9f65b664822e600"} Oct 11 04:10:30 crc kubenswrapper[4967]: I1011 04:10:30.690159 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:10:30 crc kubenswrapper[4967]: I1011 04:10:30.825239 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-r4qbf\" (UniqueName: \"kubernetes.io/projected/1317501c-8db9-4f03-a1cf-10e92b2ccc8b-kube-api-access-r4qbf\") pod \"1317501c-8db9-4f03-a1cf-10e92b2ccc8b\" (UID: \"1317501c-8db9-4f03-a1cf-10e92b2ccc8b\") " Oct 11 04:10:30 crc kubenswrapper[4967]: I1011 04:10:30.825281 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1317501c-8db9-4f03-a1cf-10e92b2ccc8b-combined-ca-bundle\") pod \"1317501c-8db9-4f03-a1cf-10e92b2ccc8b\" (UID: \"1317501c-8db9-4f03-a1cf-10e92b2ccc8b\") " Oct 11 04:10:30 crc kubenswrapper[4967]: I1011 04:10:30.825341 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1317501c-8db9-4f03-a1cf-10e92b2ccc8b-config-data\") pod \"1317501c-8db9-4f03-a1cf-10e92b2ccc8b\" (UID: \"1317501c-8db9-4f03-a1cf-10e92b2ccc8b\") " Oct 11 04:10:30 crc kubenswrapper[4967]: I1011 04:10:30.835246 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1317501c-8db9-4f03-a1cf-10e92b2ccc8b-kube-api-access-r4qbf" (OuterVolumeSpecName: "kube-api-access-r4qbf") pod "1317501c-8db9-4f03-a1cf-10e92b2ccc8b" (UID: "1317501c-8db9-4f03-a1cf-10e92b2ccc8b"). InnerVolumeSpecName "kube-api-access-r4qbf". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:10:30 crc kubenswrapper[4967]: I1011 04:10:30.857176 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1317501c-8db9-4f03-a1cf-10e92b2ccc8b-config-data" (OuterVolumeSpecName: "config-data") pod "1317501c-8db9-4f03-a1cf-10e92b2ccc8b" (UID: "1317501c-8db9-4f03-a1cf-10e92b2ccc8b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:30 crc kubenswrapper[4967]: I1011 04:10:30.859442 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1317501c-8db9-4f03-a1cf-10e92b2ccc8b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "1317501c-8db9-4f03-a1cf-10e92b2ccc8b" (UID: "1317501c-8db9-4f03-a1cf-10e92b2ccc8b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:30 crc kubenswrapper[4967]: I1011 04:10:30.927730 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-r4qbf\" (UniqueName: \"kubernetes.io/projected/1317501c-8db9-4f03-a1cf-10e92b2ccc8b-kube-api-access-r4qbf\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:30 crc kubenswrapper[4967]: I1011 04:10:30.927775 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/1317501c-8db9-4f03-a1cf-10e92b2ccc8b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:30 crc kubenswrapper[4967]: I1011 04:10:30.927788 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/1317501c-8db9-4f03-a1cf-10e92b2ccc8b-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:31 crc kubenswrapper[4967]: I1011 04:10:31.570884 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"1317501c-8db9-4f03-a1cf-10e92b2ccc8b","Type":"ContainerDied","Data":"9240fd6ade19d6577d0fda15b5465bcf854a613fc755c893f4a325590d9a1523"} Oct 11 04:10:31 crc kubenswrapper[4967]: I1011 04:10:31.570938 4967 scope.go:117] "RemoveContainer" containerID="fb17607f6184c47fb87463585b50b5bac67a191f9c083416e9f65b664822e600" Oct 11 04:10:31 crc kubenswrapper[4967]: I1011 04:10:31.571005 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:10:31 crc kubenswrapper[4967]: I1011 04:10:31.629824 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 11 04:10:31 crc kubenswrapper[4967]: I1011 04:10:31.646812 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 11 04:10:31 crc kubenswrapper[4967]: I1011 04:10:31.661595 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 11 04:10:31 crc kubenswrapper[4967]: E1011 04:10:31.661950 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="1317501c-8db9-4f03-a1cf-10e92b2ccc8b" containerName="nova-cell1-novncproxy-novncproxy" Oct 11 04:10:31 crc kubenswrapper[4967]: I1011 04:10:31.661961 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="1317501c-8db9-4f03-a1cf-10e92b2ccc8b" containerName="nova-cell1-novncproxy-novncproxy" Oct 11 04:10:31 crc kubenswrapper[4967]: I1011 04:10:31.662180 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="1317501c-8db9-4f03-a1cf-10e92b2ccc8b" containerName="nova-cell1-novncproxy-novncproxy" Oct 11 04:10:31 crc kubenswrapper[4967]: I1011 04:10:31.676444 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 11 04:10:31 crc kubenswrapper[4967]: I1011 04:10:31.676569 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:10:31 crc kubenswrapper[4967]: I1011 04:10:31.680028 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-public-svc" Oct 11 04:10:31 crc kubenswrapper[4967]: I1011 04:10:31.680258 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-novncproxy-cell1-vencrypt" Oct 11 04:10:31 crc kubenswrapper[4967]: I1011 04:10:31.681136 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-novncproxy-config-data" Oct 11 04:10:31 crc kubenswrapper[4967]: I1011 04:10:31.843568 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2914dd42-ea01-44b5-a8cd-c23fc07e3170-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"2914dd42-ea01-44b5-a8cd-c23fc07e3170\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:10:31 crc kubenswrapper[4967]: I1011 04:10:31.843719 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/2914dd42-ea01-44b5-a8cd-c23fc07e3170-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"2914dd42-ea01-44b5-a8cd-c23fc07e3170\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:10:31 crc kubenswrapper[4967]: I1011 04:10:31.843848 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5b84c\" (UniqueName: \"kubernetes.io/projected/2914dd42-ea01-44b5-a8cd-c23fc07e3170-kube-api-access-5b84c\") pod \"nova-cell1-novncproxy-0\" (UID: \"2914dd42-ea01-44b5-a8cd-c23fc07e3170\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:10:31 crc kubenswrapper[4967]: I1011 04:10:31.844062 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2914dd42-ea01-44b5-a8cd-c23fc07e3170-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"2914dd42-ea01-44b5-a8cd-c23fc07e3170\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:10:31 crc kubenswrapper[4967]: I1011 04:10:31.844194 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/2914dd42-ea01-44b5-a8cd-c23fc07e3170-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"2914dd42-ea01-44b5-a8cd-c23fc07e3170\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:10:31 crc kubenswrapper[4967]: I1011 04:10:31.946400 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/2914dd42-ea01-44b5-a8cd-c23fc07e3170-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"2914dd42-ea01-44b5-a8cd-c23fc07e3170\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:10:31 crc kubenswrapper[4967]: I1011 04:10:31.946502 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2914dd42-ea01-44b5-a8cd-c23fc07e3170-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"2914dd42-ea01-44b5-a8cd-c23fc07e3170\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:10:31 crc kubenswrapper[4967]: I1011 04:10:31.946523 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/2914dd42-ea01-44b5-a8cd-c23fc07e3170-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"2914dd42-ea01-44b5-a8cd-c23fc07e3170\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:10:31 crc kubenswrapper[4967]: I1011 04:10:31.946574 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5b84c\" (UniqueName: \"kubernetes.io/projected/2914dd42-ea01-44b5-a8cd-c23fc07e3170-kube-api-access-5b84c\") pod \"nova-cell1-novncproxy-0\" (UID: \"2914dd42-ea01-44b5-a8cd-c23fc07e3170\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:10:31 crc kubenswrapper[4967]: I1011 04:10:31.946747 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2914dd42-ea01-44b5-a8cd-c23fc07e3170-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"2914dd42-ea01-44b5-a8cd-c23fc07e3170\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:10:31 crc kubenswrapper[4967]: I1011 04:10:31.950220 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-novncproxy-tls-certs\" (UniqueName: \"kubernetes.io/secret/2914dd42-ea01-44b5-a8cd-c23fc07e3170-nova-novncproxy-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"2914dd42-ea01-44b5-a8cd-c23fc07e3170\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:10:31 crc kubenswrapper[4967]: I1011 04:10:31.950510 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/2914dd42-ea01-44b5-a8cd-c23fc07e3170-config-data\") pod \"nova-cell1-novncproxy-0\" (UID: \"2914dd42-ea01-44b5-a8cd-c23fc07e3170\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:10:31 crc kubenswrapper[4967]: I1011 04:10:31.951121 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"vencrypt-tls-certs\" (UniqueName: \"kubernetes.io/secret/2914dd42-ea01-44b5-a8cd-c23fc07e3170-vencrypt-tls-certs\") pod \"nova-cell1-novncproxy-0\" (UID: \"2914dd42-ea01-44b5-a8cd-c23fc07e3170\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:10:31 crc kubenswrapper[4967]: I1011 04:10:31.951266 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/2914dd42-ea01-44b5-a8cd-c23fc07e3170-combined-ca-bundle\") pod \"nova-cell1-novncproxy-0\" (UID: \"2914dd42-ea01-44b5-a8cd-c23fc07e3170\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:10:31 crc kubenswrapper[4967]: I1011 04:10:31.966778 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5b84c\" (UniqueName: \"kubernetes.io/projected/2914dd42-ea01-44b5-a8cd-c23fc07e3170-kube-api-access-5b84c\") pod \"nova-cell1-novncproxy-0\" (UID: \"2914dd42-ea01-44b5-a8cd-c23fc07e3170\") " pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:10:32 crc kubenswrapper[4967]: I1011 04:10:32.006474 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:10:32 crc kubenswrapper[4967]: I1011 04:10:32.488956 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-novncproxy-0"] Oct 11 04:10:32 crc kubenswrapper[4967]: W1011 04:10:32.489169 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod2914dd42_ea01_44b5_a8cd_c23fc07e3170.slice/crio-9dc77815d0a6d828ca439d3ac2ee6806dbc49c8522c9340ce5732ef0d6109ee8 WatchSource:0}: Error finding container 9dc77815d0a6d828ca439d3ac2ee6806dbc49c8522c9340ce5732ef0d6109ee8: Status 404 returned error can't find the container with id 9dc77815d0a6d828ca439d3ac2ee6806dbc49c8522c9340ce5732ef0d6109ee8 Oct 11 04:10:32 crc kubenswrapper[4967]: E1011 04:10:32.549758 4967 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1317501c_8db9_4f03_a1cf_10e92b2ccc8b.slice/crio-fb17607f6184c47fb87463585b50b5bac67a191f9c083416e9f65b664822e600.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e95b396_4e67_4df5_a39a_15ff3a00e6a1.slice/crio-f7362d856d4a38869ca9e64c37955f1a0544ca77568782afb13d718ba74b351a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e95b396_4e67_4df5_a39a_15ff3a00e6a1.slice/crio-7da08d64f43a6586cbd116ace09f566fd2c959ac5ecddb850663fc0be41dc2f2.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1317501c_8db9_4f03_a1cf_10e92b2ccc8b.slice/crio-9240fd6ade19d6577d0fda15b5465bcf854a613fc755c893f4a325590d9a1523\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83d7fc84_7338_4ad6_bf9b_901fb22a0f11.slice/crio-fce5a60c173cabbdb2679451ca90fa09a906e0f73b3616f8f59d78574a7b70d6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1317501c_8db9_4f03_a1cf_10e92b2ccc8b.slice\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83d7fc84_7338_4ad6_bf9b_901fb22a0f11.slice/crio-conmon-fce5a60c173cabbdb2679451ca90fa09a906e0f73b3616f8f59d78574a7b70d6.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83d7fc84_7338_4ad6_bf9b_901fb22a0f11.slice/crio-d82b5d8986efc77f8f88dc4bda177c119d00d1824a5b9b80a3dbdff0cbcbdde9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e95b396_4e67_4df5_a39a_15ff3a00e6a1.slice/crio-conmon-f7362d856d4a38869ca9e64c37955f1a0544ca77568782afb13d718ba74b351a.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e95b396_4e67_4df5_a39a_15ff3a00e6a1.slice/crio-conmon-d8cd6304732e32865c9d7f503d2a7316b66b5d1512696297666da84a07823c77.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod83d7fc84_7338_4ad6_bf9b_901fb22a0f11.slice/crio-conmon-d82b5d8986efc77f8f88dc4bda177c119d00d1824a5b9b80a3dbdff0cbcbdde9.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod7fd48f8c_e380_44cb_8c5b_8380b637e0ef.slice/crio-conmon-4726e4a657dc964b9f92f99fa450dccebcb517bf23d903b7075ee3e034e8ccd1.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod1317501c_8db9_4f03_a1cf_10e92b2ccc8b.slice/crio-conmon-fb17607f6184c47fb87463585b50b5bac67a191f9c083416e9f65b664822e600.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e95b396_4e67_4df5_a39a_15ff3a00e6a1.slice/crio-d8cd6304732e32865c9d7f503d2a7316b66b5d1512696297666da84a07823c77.scope\": RecentStats: unable to find data in memory cache], [\"/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod3e95b396_4e67_4df5_a39a_15ff3a00e6a1.slice/crio-conmon-7da08d64f43a6586cbd116ace09f566fd2c959ac5ecddb850663fc0be41dc2f2.scope\": RecentStats: unable to find data in memory cache]" Oct 11 04:10:32 crc kubenswrapper[4967]: I1011 04:10:32.584544 4967 generic.go:334] "Generic (PLEG): container finished" podID="7fd48f8c-e380-44cb-8c5b-8380b637e0ef" containerID="4726e4a657dc964b9f92f99fa450dccebcb517bf23d903b7075ee3e034e8ccd1" exitCode=137 Oct 11 04:10:32 crc kubenswrapper[4967]: I1011 04:10:32.584681 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7fd48f8c-e380-44cb-8c5b-8380b637e0ef","Type":"ContainerDied","Data":"4726e4a657dc964b9f92f99fa450dccebcb517bf23d903b7075ee3e034e8ccd1"} Oct 11 04:10:32 crc kubenswrapper[4967]: I1011 04:10:32.586757 4967 generic.go:334] "Generic (PLEG): container finished" podID="83d7fc84-7338-4ad6-bf9b-901fb22a0f11" containerID="fce5a60c173cabbdb2679451ca90fa09a906e0f73b3616f8f59d78574a7b70d6" exitCode=0 Oct 11 04:10:32 crc kubenswrapper[4967]: I1011 04:10:32.586815 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"83d7fc84-7338-4ad6-bf9b-901fb22a0f11","Type":"ContainerDied","Data":"fce5a60c173cabbdb2679451ca90fa09a906e0f73b3616f8f59d78574a7b70d6"} Oct 11 04:10:32 crc kubenswrapper[4967]: I1011 04:10:32.591154 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"2914dd42-ea01-44b5-a8cd-c23fc07e3170","Type":"ContainerStarted","Data":"9dc77815d0a6d828ca439d3ac2ee6806dbc49c8522c9340ce5732ef0d6109ee8"} Oct 11 04:10:32 crc kubenswrapper[4967]: I1011 04:10:32.661372 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 11 04:10:32 crc kubenswrapper[4967]: I1011 04:10:32.763434 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ggr79\" (UniqueName: \"kubernetes.io/projected/7fd48f8c-e380-44cb-8c5b-8380b637e0ef-kube-api-access-ggr79\") pod \"7fd48f8c-e380-44cb-8c5b-8380b637e0ef\" (UID: \"7fd48f8c-e380-44cb-8c5b-8380b637e0ef\") " Oct 11 04:10:32 crc kubenswrapper[4967]: I1011 04:10:32.763556 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fd48f8c-e380-44cb-8c5b-8380b637e0ef-combined-ca-bundle\") pod \"7fd48f8c-e380-44cb-8c5b-8380b637e0ef\" (UID: \"7fd48f8c-e380-44cb-8c5b-8380b637e0ef\") " Oct 11 04:10:32 crc kubenswrapper[4967]: I1011 04:10:32.763631 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fd48f8c-e380-44cb-8c5b-8380b637e0ef-config-data\") pod \"7fd48f8c-e380-44cb-8c5b-8380b637e0ef\" (UID: \"7fd48f8c-e380-44cb-8c5b-8380b637e0ef\") " Oct 11 04:10:32 crc kubenswrapper[4967]: I1011 04:10:32.771714 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7fd48f8c-e380-44cb-8c5b-8380b637e0ef-kube-api-access-ggr79" (OuterVolumeSpecName: "kube-api-access-ggr79") pod "7fd48f8c-e380-44cb-8c5b-8380b637e0ef" (UID: "7fd48f8c-e380-44cb-8c5b-8380b637e0ef"). InnerVolumeSpecName "kube-api-access-ggr79". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:10:32 crc kubenswrapper[4967]: I1011 04:10:32.793339 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fd48f8c-e380-44cb-8c5b-8380b637e0ef-config-data" (OuterVolumeSpecName: "config-data") pod "7fd48f8c-e380-44cb-8c5b-8380b637e0ef" (UID: "7fd48f8c-e380-44cb-8c5b-8380b637e0ef"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:32 crc kubenswrapper[4967]: I1011 04:10:32.793825 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7fd48f8c-e380-44cb-8c5b-8380b637e0ef-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "7fd48f8c-e380-44cb-8c5b-8380b637e0ef" (UID: "7fd48f8c-e380-44cb-8c5b-8380b637e0ef"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:32 crc kubenswrapper[4967]: I1011 04:10:32.827773 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1317501c-8db9-4f03-a1cf-10e92b2ccc8b" path="/var/lib/kubelet/pods/1317501c-8db9-4f03-a1cf-10e92b2ccc8b/volumes" Oct 11 04:10:32 crc kubenswrapper[4967]: I1011 04:10:32.854277 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 11 04:10:32 crc kubenswrapper[4967]: I1011 04:10:32.865506 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ggr79\" (UniqueName: \"kubernetes.io/projected/7fd48f8c-e380-44cb-8c5b-8380b637e0ef-kube-api-access-ggr79\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:32 crc kubenswrapper[4967]: I1011 04:10:32.865564 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/7fd48f8c-e380-44cb-8c5b-8380b637e0ef-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:32 crc kubenswrapper[4967]: I1011 04:10:32.865578 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/7fd48f8c-e380-44cb-8c5b-8380b637e0ef-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:32 crc kubenswrapper[4967]: I1011 04:10:32.968635 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83d7fc84-7338-4ad6-bf9b-901fb22a0f11-logs\") pod \"83d7fc84-7338-4ad6-bf9b-901fb22a0f11\" (UID: \"83d7fc84-7338-4ad6-bf9b-901fb22a0f11\") " Oct 11 04:10:32 crc kubenswrapper[4967]: I1011 04:10:32.968710 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tscxj\" (UniqueName: \"kubernetes.io/projected/83d7fc84-7338-4ad6-bf9b-901fb22a0f11-kube-api-access-tscxj\") pod \"83d7fc84-7338-4ad6-bf9b-901fb22a0f11\" (UID: \"83d7fc84-7338-4ad6-bf9b-901fb22a0f11\") " Oct 11 04:10:32 crc kubenswrapper[4967]: I1011 04:10:32.968792 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83d7fc84-7338-4ad6-bf9b-901fb22a0f11-combined-ca-bundle\") pod \"83d7fc84-7338-4ad6-bf9b-901fb22a0f11\" (UID: \"83d7fc84-7338-4ad6-bf9b-901fb22a0f11\") " Oct 11 04:10:32 crc kubenswrapper[4967]: I1011 04:10:32.969941 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83d7fc84-7338-4ad6-bf9b-901fb22a0f11-config-data\") pod \"83d7fc84-7338-4ad6-bf9b-901fb22a0f11\" (UID: \"83d7fc84-7338-4ad6-bf9b-901fb22a0f11\") " Oct 11 04:10:32 crc kubenswrapper[4967]: I1011 04:10:32.969776 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/83d7fc84-7338-4ad6-bf9b-901fb22a0f11-logs" (OuterVolumeSpecName: "logs") pod "83d7fc84-7338-4ad6-bf9b-901fb22a0f11" (UID: "83d7fc84-7338-4ad6-bf9b-901fb22a0f11"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:10:32 crc kubenswrapper[4967]: I1011 04:10:32.971507 4967 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/83d7fc84-7338-4ad6-bf9b-901fb22a0f11-logs\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.003120 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83d7fc84-7338-4ad6-bf9b-901fb22a0f11-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "83d7fc84-7338-4ad6-bf9b-901fb22a0f11" (UID: "83d7fc84-7338-4ad6-bf9b-901fb22a0f11"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.014453 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/83d7fc84-7338-4ad6-bf9b-901fb22a0f11-kube-api-access-tscxj" (OuterVolumeSpecName: "kube-api-access-tscxj") pod "83d7fc84-7338-4ad6-bf9b-901fb22a0f11" (UID: "83d7fc84-7338-4ad6-bf9b-901fb22a0f11"). InnerVolumeSpecName "kube-api-access-tscxj". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.025528 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/83d7fc84-7338-4ad6-bf9b-901fb22a0f11-config-data" (OuterVolumeSpecName: "config-data") pod "83d7fc84-7338-4ad6-bf9b-901fb22a0f11" (UID: "83d7fc84-7338-4ad6-bf9b-901fb22a0f11"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.072788 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/83d7fc84-7338-4ad6-bf9b-901fb22a0f11-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.073018 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/83d7fc84-7338-4ad6-bf9b-901fb22a0f11-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.073027 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tscxj\" (UniqueName: \"kubernetes.io/projected/83d7fc84-7338-4ad6-bf9b-901fb22a0f11-kube-api-access-tscxj\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.387863 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.479726 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-log-httpd\") pod \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\" (UID: \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\") " Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.480616 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-log-httpd" (OuterVolumeSpecName: "log-httpd") pod "3e95b396-4e67-4df5-a39a-15ff3a00e6a1" (UID: "3e95b396-4e67-4df5-a39a-15ff3a00e6a1"). InnerVolumeSpecName "log-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.480693 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-run-httpd\") pod \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\" (UID: \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\") " Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.481009 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-run-httpd" (OuterVolumeSpecName: "run-httpd") pod "3e95b396-4e67-4df5-a39a-15ff3a00e6a1" (UID: "3e95b396-4e67-4df5-a39a-15ff3a00e6a1"). InnerVolumeSpecName "run-httpd". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.481083 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-config-data\") pod \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\" (UID: \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\") " Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.481109 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-plj5z\" (UniqueName: \"kubernetes.io/projected/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-kube-api-access-plj5z\") pod \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\" (UID: \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\") " Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.481142 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-sg-core-conf-yaml\") pod \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\" (UID: \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\") " Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.481251 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-combined-ca-bundle\") pod \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\" (UID: \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\") " Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.481305 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-ceilometer-tls-certs\") pod \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\" (UID: \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\") " Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.481328 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-scripts\") pod \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\" (UID: \"3e95b396-4e67-4df5-a39a-15ff3a00e6a1\") " Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.482337 4967 reconciler_common.go:293] "Volume detached for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-log-httpd\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.482358 4967 reconciler_common.go:293] "Volume detached for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-run-httpd\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.487877 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-scripts" (OuterVolumeSpecName: "scripts") pod "3e95b396-4e67-4df5-a39a-15ff3a00e6a1" (UID: "3e95b396-4e67-4df5-a39a-15ff3a00e6a1"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.490267 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-kube-api-access-plj5z" (OuterVolumeSpecName: "kube-api-access-plj5z") pod "3e95b396-4e67-4df5-a39a-15ff3a00e6a1" (UID: "3e95b396-4e67-4df5-a39a-15ff3a00e6a1"). InnerVolumeSpecName "kube-api-access-plj5z". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.520678 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-sg-core-conf-yaml" (OuterVolumeSpecName: "sg-core-conf-yaml") pod "3e95b396-4e67-4df5-a39a-15ff3a00e6a1" (UID: "3e95b396-4e67-4df5-a39a-15ff3a00e6a1"). InnerVolumeSpecName "sg-core-conf-yaml". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.560613 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-ceilometer-tls-certs" (OuterVolumeSpecName: "ceilometer-tls-certs") pod "3e95b396-4e67-4df5-a39a-15ff3a00e6a1" (UID: "3e95b396-4e67-4df5-a39a-15ff3a00e6a1"). InnerVolumeSpecName "ceilometer-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.584436 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.584476 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-plj5z\" (UniqueName: \"kubernetes.io/projected/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-kube-api-access-plj5z\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.584490 4967 reconciler_common.go:293] "Volume detached for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-sg-core-conf-yaml\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.584501 4967 reconciler_common.go:293] "Volume detached for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-ceilometer-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.586311 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "3e95b396-4e67-4df5-a39a-15ff3a00e6a1" (UID: "3e95b396-4e67-4df5-a39a-15ff3a00e6a1"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.610773 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-config-data" (OuterVolumeSpecName: "config-data") pod "3e95b396-4e67-4df5-a39a-15ff3a00e6a1" (UID: "3e95b396-4e67-4df5-a39a-15ff3a00e6a1"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.619480 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.619622 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"7fd48f8c-e380-44cb-8c5b-8380b637e0ef","Type":"ContainerDied","Data":"2ae1533565367937fc2b5289cbea87e4c92e173b211be574ab405dfd4e8cdf7e"} Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.619729 4967 scope.go:117] "RemoveContainer" containerID="4726e4a657dc964b9f92f99fa450dccebcb517bf23d903b7075ee3e034e8ccd1" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.626864 4967 generic.go:334] "Generic (PLEG): container finished" podID="3e95b396-4e67-4df5-a39a-15ff3a00e6a1" containerID="0875cf7cbc65df1e3e6e99f37987fe6f36adf3677c200154187ee9d1072fac73" exitCode=0 Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.627489 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e95b396-4e67-4df5-a39a-15ff3a00e6a1","Type":"ContainerDied","Data":"0875cf7cbc65df1e3e6e99f37987fe6f36adf3677c200154187ee9d1072fac73"} Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.627557 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"3e95b396-4e67-4df5-a39a-15ff3a00e6a1","Type":"ContainerDied","Data":"e226bc76f956e91007e49bc6d6a7b72a16216c26fa9caa5ec70ba33caa30e164"} Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.628251 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.633266 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"83d7fc84-7338-4ad6-bf9b-901fb22a0f11","Type":"ContainerDied","Data":"fc72043565edcb5f9f5a6bf9ed11a65f76939b465152281cd8555570f944dfbd"} Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.633301 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.635622 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-novncproxy-0" event={"ID":"2914dd42-ea01-44b5-a8cd-c23fc07e3170","Type":"ContainerStarted","Data":"649b52a83043912ff083b31b4d85073eb0f8d1e6f22b469aa596015c9e6e8c09"} Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.651110 4967 scope.go:117] "RemoveContainer" containerID="7da08d64f43a6586cbd116ace09f566fd2c959ac5ecddb850663fc0be41dc2f2" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.660179 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.679222 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.685645 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.685670 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/3e95b396-4e67-4df5-a39a-15ff3a00e6a1-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.696705 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 11 04:10:33 crc kubenswrapper[4967]: E1011 04:10:33.712745 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e95b396-4e67-4df5-a39a-15ff3a00e6a1" containerName="sg-core" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.712800 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e95b396-4e67-4df5-a39a-15ff3a00e6a1" containerName="sg-core" Oct 11 04:10:33 crc kubenswrapper[4967]: E1011 04:10:33.712835 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7fd48f8c-e380-44cb-8c5b-8380b637e0ef" containerName="nova-scheduler-scheduler" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.712843 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="7fd48f8c-e380-44cb-8c5b-8380b637e0ef" containerName="nova-scheduler-scheduler" Oct 11 04:10:33 crc kubenswrapper[4967]: E1011 04:10:33.712866 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e95b396-4e67-4df5-a39a-15ff3a00e6a1" containerName="ceilometer-central-agent" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.712876 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e95b396-4e67-4df5-a39a-15ff3a00e6a1" containerName="ceilometer-central-agent" Oct 11 04:10:33 crc kubenswrapper[4967]: E1011 04:10:33.712894 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e95b396-4e67-4df5-a39a-15ff3a00e6a1" containerName="proxy-httpd" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.712900 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e95b396-4e67-4df5-a39a-15ff3a00e6a1" containerName="proxy-httpd" Oct 11 04:10:33 crc kubenswrapper[4967]: E1011 04:10:33.712911 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3e95b396-4e67-4df5-a39a-15ff3a00e6a1" containerName="ceilometer-notification-agent" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.712916 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="3e95b396-4e67-4df5-a39a-15ff3a00e6a1" containerName="ceilometer-notification-agent" Oct 11 04:10:33 crc kubenswrapper[4967]: E1011 04:10:33.712947 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83d7fc84-7338-4ad6-bf9b-901fb22a0f11" containerName="nova-api-log" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.712955 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="83d7fc84-7338-4ad6-bf9b-901fb22a0f11" containerName="nova-api-log" Oct 11 04:10:33 crc kubenswrapper[4967]: E1011 04:10:33.712983 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="83d7fc84-7338-4ad6-bf9b-901fb22a0f11" containerName="nova-api-api" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.712992 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="83d7fc84-7338-4ad6-bf9b-901fb22a0f11" containerName="nova-api-api" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.713614 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="7fd48f8c-e380-44cb-8c5b-8380b637e0ef" containerName="nova-scheduler-scheduler" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.713642 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e95b396-4e67-4df5-a39a-15ff3a00e6a1" containerName="proxy-httpd" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.713655 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="83d7fc84-7338-4ad6-bf9b-901fb22a0f11" containerName="nova-api-api" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.713676 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e95b396-4e67-4df5-a39a-15ff3a00e6a1" containerName="ceilometer-notification-agent" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.713684 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e95b396-4e67-4df5-a39a-15ff3a00e6a1" containerName="sg-core" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.713693 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="83d7fc84-7338-4ad6-bf9b-901fb22a0f11" containerName="nova-api-log" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.713716 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="3e95b396-4e67-4df5-a39a-15ff3a00e6a1" containerName="ceilometer-central-agent" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.714638 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.714747 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.715873 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-novncproxy-0" podStartSLOduration=2.715852459 podStartE2EDuration="2.715852459s" podCreationTimestamp="2025-10-11 04:10:31 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:10:33.670688373 +0000 UTC m=+1161.633897306" watchObservedRunningTime="2025-10-11 04:10:33.715852459 +0000 UTC m=+1161.679061392" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.721662 4967 scope.go:117] "RemoveContainer" containerID="d8cd6304732e32865c9d7f503d2a7316b66b5d1512696297666da84a07823c77" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.733156 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.775211 4967 scope.go:117] "RemoveContainer" containerID="0875cf7cbc65df1e3e6e99f37987fe6f36adf3677c200154187ee9d1072fac73" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.787611 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecf46e75-f69c-4049-ada3-f74914ae8346-config-data\") pod \"nova-scheduler-0\" (UID: \"ecf46e75-f69c-4049-ada3-f74914ae8346\") " pod="openstack/nova-scheduler-0" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.787655 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fsqmd\" (UniqueName: \"kubernetes.io/projected/ecf46e75-f69c-4049-ada3-f74914ae8346-kube-api-access-fsqmd\") pod \"nova-scheduler-0\" (UID: \"ecf46e75-f69c-4049-ada3-f74914ae8346\") " pod="openstack/nova-scheduler-0" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.787759 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecf46e75-f69c-4049-ada3-f74914ae8346-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ecf46e75-f69c-4049-ada3-f74914ae8346\") " pod="openstack/nova-scheduler-0" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.788853 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.807567 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.820317 4967 scope.go:117] "RemoveContainer" containerID="f7362d856d4a38869ca9e64c37955f1a0544ca77568782afb13d718ba74b351a" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.832804 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.857141 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.860231 4967 scope.go:117] "RemoveContainer" containerID="7da08d64f43a6586cbd116ace09f566fd2c959ac5ecddb850663fc0be41dc2f2" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.860417 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 11 04:10:33 crc kubenswrapper[4967]: E1011 04:10:33.861023 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7da08d64f43a6586cbd116ace09f566fd2c959ac5ecddb850663fc0be41dc2f2\": container with ID starting with 7da08d64f43a6586cbd116ace09f566fd2c959ac5ecddb850663fc0be41dc2f2 not found: ID does not exist" containerID="7da08d64f43a6586cbd116ace09f566fd2c959ac5ecddb850663fc0be41dc2f2" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.861414 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7da08d64f43a6586cbd116ace09f566fd2c959ac5ecddb850663fc0be41dc2f2"} err="failed to get container status \"7da08d64f43a6586cbd116ace09f566fd2c959ac5ecddb850663fc0be41dc2f2\": rpc error: code = NotFound desc = could not find container \"7da08d64f43a6586cbd116ace09f566fd2c959ac5ecddb850663fc0be41dc2f2\": container with ID starting with 7da08d64f43a6586cbd116ace09f566fd2c959ac5ecddb850663fc0be41dc2f2 not found: ID does not exist" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.861448 4967 scope.go:117] "RemoveContainer" containerID="d8cd6304732e32865c9d7f503d2a7316b66b5d1512696297666da84a07823c77" Oct 11 04:10:33 crc kubenswrapper[4967]: E1011 04:10:33.867690 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d8cd6304732e32865c9d7f503d2a7316b66b5d1512696297666da84a07823c77\": container with ID starting with d8cd6304732e32865c9d7f503d2a7316b66b5d1512696297666da84a07823c77 not found: ID does not exist" containerID="d8cd6304732e32865c9d7f503d2a7316b66b5d1512696297666da84a07823c77" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.867746 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d8cd6304732e32865c9d7f503d2a7316b66b5d1512696297666da84a07823c77"} err="failed to get container status \"d8cd6304732e32865c9d7f503d2a7316b66b5d1512696297666da84a07823c77\": rpc error: code = NotFound desc = could not find container \"d8cd6304732e32865c9d7f503d2a7316b66b5d1512696297666da84a07823c77\": container with ID starting with d8cd6304732e32865c9d7f503d2a7316b66b5d1512696297666da84a07823c77 not found: ID does not exist" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.867775 4967 scope.go:117] "RemoveContainer" containerID="0875cf7cbc65df1e3e6e99f37987fe6f36adf3677c200154187ee9d1072fac73" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.867995 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.868059 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.868245 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 11 04:10:33 crc kubenswrapper[4967]: E1011 04:10:33.868397 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0875cf7cbc65df1e3e6e99f37987fe6f36adf3677c200154187ee9d1072fac73\": container with ID starting with 0875cf7cbc65df1e3e6e99f37987fe6f36adf3677c200154187ee9d1072fac73 not found: ID does not exist" containerID="0875cf7cbc65df1e3e6e99f37987fe6f36adf3677c200154187ee9d1072fac73" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.868424 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0875cf7cbc65df1e3e6e99f37987fe6f36adf3677c200154187ee9d1072fac73"} err="failed to get container status \"0875cf7cbc65df1e3e6e99f37987fe6f36adf3677c200154187ee9d1072fac73\": rpc error: code = NotFound desc = could not find container \"0875cf7cbc65df1e3e6e99f37987fe6f36adf3677c200154187ee9d1072fac73\": container with ID starting with 0875cf7cbc65df1e3e6e99f37987fe6f36adf3677c200154187ee9d1072fac73 not found: ID does not exist" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.868449 4967 scope.go:117] "RemoveContainer" containerID="f7362d856d4a38869ca9e64c37955f1a0544ca77568782afb13d718ba74b351a" Oct 11 04:10:33 crc kubenswrapper[4967]: E1011 04:10:33.868755 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f7362d856d4a38869ca9e64c37955f1a0544ca77568782afb13d718ba74b351a\": container with ID starting with f7362d856d4a38869ca9e64c37955f1a0544ca77568782afb13d718ba74b351a not found: ID does not exist" containerID="f7362d856d4a38869ca9e64c37955f1a0544ca77568782afb13d718ba74b351a" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.868776 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f7362d856d4a38869ca9e64c37955f1a0544ca77568782afb13d718ba74b351a"} err="failed to get container status \"f7362d856d4a38869ca9e64c37955f1a0544ca77568782afb13d718ba74b351a\": rpc error: code = NotFound desc = could not find container \"f7362d856d4a38869ca9e64c37955f1a0544ca77568782afb13d718ba74b351a\": container with ID starting with f7362d856d4a38869ca9e64c37955f1a0544ca77568782afb13d718ba74b351a not found: ID does not exist" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.868788 4967 scope.go:117] "RemoveContainer" containerID="fce5a60c173cabbdb2679451ca90fa09a906e0f73b3616f8f59d78574a7b70d6" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.871687 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.884421 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.890195 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecf46e75-f69c-4049-ada3-f74914ae8346-config-data\") pod \"nova-scheduler-0\" (UID: \"ecf46e75-f69c-4049-ada3-f74914ae8346\") " pod="openstack/nova-scheduler-0" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.890246 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fsqmd\" (UniqueName: \"kubernetes.io/projected/ecf46e75-f69c-4049-ada3-f74914ae8346-kube-api-access-fsqmd\") pod \"nova-scheduler-0\" (UID: \"ecf46e75-f69c-4049-ada3-f74914ae8346\") " pod="openstack/nova-scheduler-0" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.890522 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecf46e75-f69c-4049-ada3-f74914ae8346-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ecf46e75-f69c-4049-ada3-f74914ae8346\") " pod="openstack/nova-scheduler-0" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.895088 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.896719 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecf46e75-f69c-4049-ada3-f74914ae8346-config-data\") pod \"nova-scheduler-0\" (UID: \"ecf46e75-f69c-4049-ada3-f74914ae8346\") " pod="openstack/nova-scheduler-0" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.898202 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.902756 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecf46e75-f69c-4049-ada3-f74914ae8346-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"ecf46e75-f69c-4049-ada3-f74914ae8346\") " pod="openstack/nova-scheduler-0" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.902822 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-config-data" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.902837 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-ceilometer-internal-svc" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.903015 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"ceilometer-scripts" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.907752 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.933588 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fsqmd\" (UniqueName: \"kubernetes.io/projected/ecf46e75-f69c-4049-ada3-f74914ae8346-kube-api-access-fsqmd\") pod \"nova-scheduler-0\" (UID: \"ecf46e75-f69c-4049-ada3-f74914ae8346\") " pod="openstack/nova-scheduler-0" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.992625 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c47c6348-b68f-410e-93d7-c10eed165a34-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c47c6348-b68f-410e-93d7-c10eed165a34\") " pod="openstack/ceilometer-0" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.992681 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7\") " pod="openstack/nova-api-0" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.992702 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c47c6348-b68f-410e-93d7-c10eed165a34-config-data\") pod \"ceilometer-0\" (UID: \"c47c6348-b68f-410e-93d7-c10eed165a34\") " pod="openstack/ceilometer-0" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.992855 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-public-tls-certs\") pod \"nova-api-0\" (UID: \"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7\") " pod="openstack/nova-api-0" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.992916 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-config-data\") pod \"nova-api-0\" (UID: \"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7\") " pod="openstack/nova-api-0" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.992971 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c47c6348-b68f-410e-93d7-c10eed165a34-scripts\") pod \"ceilometer-0\" (UID: \"c47c6348-b68f-410e-93d7-c10eed165a34\") " pod="openstack/ceilometer-0" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.993000 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ck5d2\" (UniqueName: \"kubernetes.io/projected/c47c6348-b68f-410e-93d7-c10eed165a34-kube-api-access-ck5d2\") pod \"ceilometer-0\" (UID: \"c47c6348-b68f-410e-93d7-c10eed165a34\") " pod="openstack/ceilometer-0" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.993026 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c47c6348-b68f-410e-93d7-c10eed165a34-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c47c6348-b68f-410e-93d7-c10eed165a34\") " pod="openstack/ceilometer-0" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.993109 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c47c6348-b68f-410e-93d7-c10eed165a34-log-httpd\") pod \"ceilometer-0\" (UID: \"c47c6348-b68f-410e-93d7-c10eed165a34\") " pod="openstack/ceilometer-0" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.993155 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c47c6348-b68f-410e-93d7-c10eed165a34-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c47c6348-b68f-410e-93d7-c10eed165a34\") " pod="openstack/ceilometer-0" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.993263 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jxkkv\" (UniqueName: \"kubernetes.io/projected/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-kube-api-access-jxkkv\") pod \"nova-api-0\" (UID: \"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7\") " pod="openstack/nova-api-0" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.993335 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7\") " pod="openstack/nova-api-0" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.993397 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c47c6348-b68f-410e-93d7-c10eed165a34-run-httpd\") pod \"ceilometer-0\" (UID: \"c47c6348-b68f-410e-93d7-c10eed165a34\") " pod="openstack/ceilometer-0" Oct 11 04:10:33 crc kubenswrapper[4967]: I1011 04:10:33.993468 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-logs\") pod \"nova-api-0\" (UID: \"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7\") " pod="openstack/nova-api-0" Oct 11 04:10:34 crc kubenswrapper[4967]: I1011 04:10:34.005812 4967 scope.go:117] "RemoveContainer" containerID="d82b5d8986efc77f8f88dc4bda177c119d00d1824a5b9b80a3dbdff0cbcbdde9" Oct 11 04:10:34 crc kubenswrapper[4967]: I1011 04:10:34.077798 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 11 04:10:34 crc kubenswrapper[4967]: I1011 04:10:34.095568 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c47c6348-b68f-410e-93d7-c10eed165a34-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c47c6348-b68f-410e-93d7-c10eed165a34\") " pod="openstack/ceilometer-0" Oct 11 04:10:34 crc kubenswrapper[4967]: I1011 04:10:34.095614 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7\") " pod="openstack/nova-api-0" Oct 11 04:10:34 crc kubenswrapper[4967]: I1011 04:10:34.095632 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c47c6348-b68f-410e-93d7-c10eed165a34-config-data\") pod \"ceilometer-0\" (UID: \"c47c6348-b68f-410e-93d7-c10eed165a34\") " pod="openstack/ceilometer-0" Oct 11 04:10:34 crc kubenswrapper[4967]: I1011 04:10:34.095666 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-public-tls-certs\") pod \"nova-api-0\" (UID: \"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7\") " pod="openstack/nova-api-0" Oct 11 04:10:34 crc kubenswrapper[4967]: I1011 04:10:34.095689 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-config-data\") pod \"nova-api-0\" (UID: \"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7\") " pod="openstack/nova-api-0" Oct 11 04:10:34 crc kubenswrapper[4967]: I1011 04:10:34.095707 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c47c6348-b68f-410e-93d7-c10eed165a34-scripts\") pod \"ceilometer-0\" (UID: \"c47c6348-b68f-410e-93d7-c10eed165a34\") " pod="openstack/ceilometer-0" Oct 11 04:10:34 crc kubenswrapper[4967]: I1011 04:10:34.095722 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ck5d2\" (UniqueName: \"kubernetes.io/projected/c47c6348-b68f-410e-93d7-c10eed165a34-kube-api-access-ck5d2\") pod \"ceilometer-0\" (UID: \"c47c6348-b68f-410e-93d7-c10eed165a34\") " pod="openstack/ceilometer-0" Oct 11 04:10:34 crc kubenswrapper[4967]: I1011 04:10:34.095738 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c47c6348-b68f-410e-93d7-c10eed165a34-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c47c6348-b68f-410e-93d7-c10eed165a34\") " pod="openstack/ceilometer-0" Oct 11 04:10:34 crc kubenswrapper[4967]: I1011 04:10:34.095762 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c47c6348-b68f-410e-93d7-c10eed165a34-log-httpd\") pod \"ceilometer-0\" (UID: \"c47c6348-b68f-410e-93d7-c10eed165a34\") " pod="openstack/ceilometer-0" Oct 11 04:10:34 crc kubenswrapper[4967]: I1011 04:10:34.095783 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c47c6348-b68f-410e-93d7-c10eed165a34-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c47c6348-b68f-410e-93d7-c10eed165a34\") " pod="openstack/ceilometer-0" Oct 11 04:10:34 crc kubenswrapper[4967]: I1011 04:10:34.095834 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jxkkv\" (UniqueName: \"kubernetes.io/projected/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-kube-api-access-jxkkv\") pod \"nova-api-0\" (UID: \"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7\") " pod="openstack/nova-api-0" Oct 11 04:10:34 crc kubenswrapper[4967]: I1011 04:10:34.095867 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7\") " pod="openstack/nova-api-0" Oct 11 04:10:34 crc kubenswrapper[4967]: I1011 04:10:34.095894 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c47c6348-b68f-410e-93d7-c10eed165a34-run-httpd\") pod \"ceilometer-0\" (UID: \"c47c6348-b68f-410e-93d7-c10eed165a34\") " pod="openstack/ceilometer-0" Oct 11 04:10:34 crc kubenswrapper[4967]: I1011 04:10:34.095919 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-logs\") pod \"nova-api-0\" (UID: \"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7\") " pod="openstack/nova-api-0" Oct 11 04:10:34 crc kubenswrapper[4967]: I1011 04:10:34.096318 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-logs\") pod \"nova-api-0\" (UID: \"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7\") " pod="openstack/nova-api-0" Oct 11 04:10:34 crc kubenswrapper[4967]: I1011 04:10:34.097621 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c47c6348-b68f-410e-93d7-c10eed165a34-log-httpd\") pod \"ceilometer-0\" (UID: \"c47c6348-b68f-410e-93d7-c10eed165a34\") " pod="openstack/ceilometer-0" Oct 11 04:10:34 crc kubenswrapper[4967]: I1011 04:10:34.100032 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/c47c6348-b68f-410e-93d7-c10eed165a34-combined-ca-bundle\") pod \"ceilometer-0\" (UID: \"c47c6348-b68f-410e-93d7-c10eed165a34\") " pod="openstack/ceilometer-0" Oct 11 04:10:34 crc kubenswrapper[4967]: I1011 04:10:34.100566 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ceilometer-tls-certs\" (UniqueName: \"kubernetes.io/secret/c47c6348-b68f-410e-93d7-c10eed165a34-ceilometer-tls-certs\") pod \"ceilometer-0\" (UID: \"c47c6348-b68f-410e-93d7-c10eed165a34\") " pod="openstack/ceilometer-0" Oct 11 04:10:34 crc kubenswrapper[4967]: I1011 04:10:34.101444 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-httpd\" (UniqueName: \"kubernetes.io/empty-dir/c47c6348-b68f-410e-93d7-c10eed165a34-run-httpd\") pod \"ceilometer-0\" (UID: \"c47c6348-b68f-410e-93d7-c10eed165a34\") " pod="openstack/ceilometer-0" Oct 11 04:10:34 crc kubenswrapper[4967]: I1011 04:10:34.102546 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"sg-core-conf-yaml\" (UniqueName: \"kubernetes.io/secret/c47c6348-b68f-410e-93d7-c10eed165a34-sg-core-conf-yaml\") pod \"ceilometer-0\" (UID: \"c47c6348-b68f-410e-93d7-c10eed165a34\") " pod="openstack/ceilometer-0" Oct 11 04:10:34 crc kubenswrapper[4967]: I1011 04:10:34.102924 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-internal-tls-certs\") pod \"nova-api-0\" (UID: \"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7\") " pod="openstack/nova-api-0" Oct 11 04:10:34 crc kubenswrapper[4967]: I1011 04:10:34.104056 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-public-tls-certs\") pod \"nova-api-0\" (UID: \"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7\") " pod="openstack/nova-api-0" Oct 11 04:10:34 crc kubenswrapper[4967]: I1011 04:10:34.106731 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/c47c6348-b68f-410e-93d7-c10eed165a34-scripts\") pod \"ceilometer-0\" (UID: \"c47c6348-b68f-410e-93d7-c10eed165a34\") " pod="openstack/ceilometer-0" Oct 11 04:10:34 crc kubenswrapper[4967]: I1011 04:10:34.107038 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-config-data\") pod \"nova-api-0\" (UID: \"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7\") " pod="openstack/nova-api-0" Oct 11 04:10:34 crc kubenswrapper[4967]: I1011 04:10:34.107894 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7\") " pod="openstack/nova-api-0" Oct 11 04:10:34 crc kubenswrapper[4967]: I1011 04:10:34.113356 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/c47c6348-b68f-410e-93d7-c10eed165a34-config-data\") pod \"ceilometer-0\" (UID: \"c47c6348-b68f-410e-93d7-c10eed165a34\") " pod="openstack/ceilometer-0" Oct 11 04:10:34 crc kubenswrapper[4967]: I1011 04:10:34.114725 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ck5d2\" (UniqueName: \"kubernetes.io/projected/c47c6348-b68f-410e-93d7-c10eed165a34-kube-api-access-ck5d2\") pod \"ceilometer-0\" (UID: \"c47c6348-b68f-410e-93d7-c10eed165a34\") " pod="openstack/ceilometer-0" Oct 11 04:10:34 crc kubenswrapper[4967]: I1011 04:10:34.123566 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jxkkv\" (UniqueName: \"kubernetes.io/projected/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-kube-api-access-jxkkv\") pod \"nova-api-0\" (UID: \"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7\") " pod="openstack/nova-api-0" Oct 11 04:10:34 crc kubenswrapper[4967]: I1011 04:10:34.307133 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 11 04:10:35 crc kubenswrapper[4967]: I1011 04:10:34.316569 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/ceilometer-0" Oct 11 04:10:35 crc kubenswrapper[4967]: I1011 04:10:34.533972 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 11 04:10:35 crc kubenswrapper[4967]: W1011 04:10:34.546746 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podecf46e75_f69c_4049_ada3_f74914ae8346.slice/crio-3dfc40a7aa0ede18b4dd70eefc3e4f423c75904366af6b65d8c96ade852176ff WatchSource:0}: Error finding container 3dfc40a7aa0ede18b4dd70eefc3e4f423c75904366af6b65d8c96ade852176ff: Status 404 returned error can't find the container with id 3dfc40a7aa0ede18b4dd70eefc3e4f423c75904366af6b65d8c96ade852176ff Oct 11 04:10:35 crc kubenswrapper[4967]: I1011 04:10:34.660553 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ecf46e75-f69c-4049-ada3-f74914ae8346","Type":"ContainerStarted","Data":"3dfc40a7aa0ede18b4dd70eefc3e4f423c75904366af6b65d8c96ade852176ff"} Oct 11 04:10:35 crc kubenswrapper[4967]: I1011 04:10:34.830793 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3e95b396-4e67-4df5-a39a-15ff3a00e6a1" path="/var/lib/kubelet/pods/3e95b396-4e67-4df5-a39a-15ff3a00e6a1/volumes" Oct 11 04:10:35 crc kubenswrapper[4967]: I1011 04:10:34.831574 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7fd48f8c-e380-44cb-8c5b-8380b637e0ef" path="/var/lib/kubelet/pods/7fd48f8c-e380-44cb-8c5b-8380b637e0ef/volumes" Oct 11 04:10:35 crc kubenswrapper[4967]: I1011 04:10:34.832144 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="83d7fc84-7338-4ad6-bf9b-901fb22a0f11" path="/var/lib/kubelet/pods/83d7fc84-7338-4ad6-bf9b-901fb22a0f11/volumes" Oct 11 04:10:35 crc kubenswrapper[4967]: I1011 04:10:35.511332 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 11 04:10:35 crc kubenswrapper[4967]: I1011 04:10:35.522632 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/ceilometer-0"] Oct 11 04:10:35 crc kubenswrapper[4967]: W1011 04:10:35.526218 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podc47c6348_b68f_410e_93d7_c10eed165a34.slice/crio-81459c0fb0a6da7b070657d6b9f898e0ed9555443b5f6a1d0fc0aed3e174bc1b WatchSource:0}: Error finding container 81459c0fb0a6da7b070657d6b9f898e0ed9555443b5f6a1d0fc0aed3e174bc1b: Status 404 returned error can't find the container with id 81459c0fb0a6da7b070657d6b9f898e0ed9555443b5f6a1d0fc0aed3e174bc1b Oct 11 04:10:35 crc kubenswrapper[4967]: I1011 04:10:35.673174 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c47c6348-b68f-410e-93d7-c10eed165a34","Type":"ContainerStarted","Data":"81459c0fb0a6da7b070657d6b9f898e0ed9555443b5f6a1d0fc0aed3e174bc1b"} Oct 11 04:10:35 crc kubenswrapper[4967]: I1011 04:10:35.676238 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ecf46e75-f69c-4049-ada3-f74914ae8346","Type":"ContainerStarted","Data":"0ff59e5bf52f194cbaeaa5baec2131cf19c3f2cb5a5617318169fb63f10618d9"} Oct 11 04:10:35 crc kubenswrapper[4967]: I1011 04:10:35.677424 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7","Type":"ContainerStarted","Data":"3dc831d1cb19848fa8ff46ed907560ccdc0c2a5a941baa673d1cc5242086fe98"} Oct 11 04:10:35 crc kubenswrapper[4967]: I1011 04:10:35.703537 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=2.7035200379999997 podStartE2EDuration="2.703520038s" podCreationTimestamp="2025-10-11 04:10:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:10:35.693578628 +0000 UTC m=+1163.656787561" watchObservedRunningTime="2025-10-11 04:10:35.703520038 +0000 UTC m=+1163.666728971" Oct 11 04:10:35 crc kubenswrapper[4967]: I1011 04:10:35.990263 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/dnsmasq-dns-cd5cbd7b9-87xhd" Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.076242 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-xbm92"] Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.076532 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/dnsmasq-dns-bccf8f775-xbm92" podUID="2a0b5b97-4605-4d4f-927f-d1eafe06956f" containerName="dnsmasq-dns" containerID="cri-o://e42a65cb81472bcfcb031eb8b1298f5038365ed72c88b70eac57614ce427dde4" gracePeriod=10 Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.508808 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-xbm92" Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.651373 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2a0b5b97-4605-4d4f-927f-d1eafe06956f-ovsdbserver-sb\") pod \"2a0b5b97-4605-4d4f-927f-d1eafe06956f\" (UID: \"2a0b5b97-4605-4d4f-927f-d1eafe06956f\") " Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.651821 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a0b5b97-4605-4d4f-927f-d1eafe06956f-config\") pod \"2a0b5b97-4605-4d4f-927f-d1eafe06956f\" (UID: \"2a0b5b97-4605-4d4f-927f-d1eafe06956f\") " Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.651941 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4rgrq\" (UniqueName: \"kubernetes.io/projected/2a0b5b97-4605-4d4f-927f-d1eafe06956f-kube-api-access-4rgrq\") pod \"2a0b5b97-4605-4d4f-927f-d1eafe06956f\" (UID: \"2a0b5b97-4605-4d4f-927f-d1eafe06956f\") " Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.652060 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a0b5b97-4605-4d4f-927f-d1eafe06956f-dns-svc\") pod \"2a0b5b97-4605-4d4f-927f-d1eafe06956f\" (UID: \"2a0b5b97-4605-4d4f-927f-d1eafe06956f\") " Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.652298 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2a0b5b97-4605-4d4f-927f-d1eafe06956f-ovsdbserver-nb\") pod \"2a0b5b97-4605-4d4f-927f-d1eafe06956f\" (UID: \"2a0b5b97-4605-4d4f-927f-d1eafe06956f\") " Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.652501 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2a0b5b97-4605-4d4f-927f-d1eafe06956f-dns-swift-storage-0\") pod \"2a0b5b97-4605-4d4f-927f-d1eafe06956f\" (UID: \"2a0b5b97-4605-4d4f-927f-d1eafe06956f\") " Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.668325 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a0b5b97-4605-4d4f-927f-d1eafe06956f-kube-api-access-4rgrq" (OuterVolumeSpecName: "kube-api-access-4rgrq") pod "2a0b5b97-4605-4d4f-927f-d1eafe06956f" (UID: "2a0b5b97-4605-4d4f-927f-d1eafe06956f"). InnerVolumeSpecName "kube-api-access-4rgrq". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.713519 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7","Type":"ContainerStarted","Data":"0656e8fb8f6f7b37d2f167a44285df4d9003297f85025352040830189eaa3c1e"} Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.713946 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7","Type":"ContainerStarted","Data":"6322bd913c3c8da7f4f00e75fa8002b5ceda98f094f842d81283ab3ae5c2f7b0"} Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.717238 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c47c6348-b68f-410e-93d7-c10eed165a34","Type":"ContainerStarted","Data":"b6e72fa43b3eb298b8cfafb2d2f34da9e7e1ff4aa440817b0db54ce7ca8bb827"} Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.720639 4967 generic.go:334] "Generic (PLEG): container finished" podID="2a0b5b97-4605-4d4f-927f-d1eafe06956f" containerID="e42a65cb81472bcfcb031eb8b1298f5038365ed72c88b70eac57614ce427dde4" exitCode=0 Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.720684 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/dnsmasq-dns-bccf8f775-xbm92" Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.720731 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-xbm92" event={"ID":"2a0b5b97-4605-4d4f-927f-d1eafe06956f","Type":"ContainerDied","Data":"e42a65cb81472bcfcb031eb8b1298f5038365ed72c88b70eac57614ce427dde4"} Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.720794 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/dnsmasq-dns-bccf8f775-xbm92" event={"ID":"2a0b5b97-4605-4d4f-927f-d1eafe06956f","Type":"ContainerDied","Data":"38de4d739ca79b2d47266353bb6177bd566656f98f4c935c1653a61487cd1e25"} Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.720821 4967 scope.go:117] "RemoveContainer" containerID="e42a65cb81472bcfcb031eb8b1298f5038365ed72c88b70eac57614ce427dde4" Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.723087 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a0b5b97-4605-4d4f-927f-d1eafe06956f-dns-swift-storage-0" (OuterVolumeSpecName: "dns-swift-storage-0") pod "2a0b5b97-4605-4d4f-927f-d1eafe06956f" (UID: "2a0b5b97-4605-4d4f-927f-d1eafe06956f"). InnerVolumeSpecName "dns-swift-storage-0". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.743632 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a0b5b97-4605-4d4f-927f-d1eafe06956f-ovsdbserver-sb" (OuterVolumeSpecName: "ovsdbserver-sb") pod "2a0b5b97-4605-4d4f-927f-d1eafe06956f" (UID: "2a0b5b97-4605-4d4f-927f-d1eafe06956f"). InnerVolumeSpecName "ovsdbserver-sb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.744194 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=3.744177708 podStartE2EDuration="3.744177708s" podCreationTimestamp="2025-10-11 04:10:33 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:10:36.739290915 +0000 UTC m=+1164.702499868" watchObservedRunningTime="2025-10-11 04:10:36.744177708 +0000 UTC m=+1164.707386641" Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.748579 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a0b5b97-4605-4d4f-927f-d1eafe06956f-config" (OuterVolumeSpecName: "config") pod "2a0b5b97-4605-4d4f-927f-d1eafe06956f" (UID: "2a0b5b97-4605-4d4f-927f-d1eafe06956f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.749773 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a0b5b97-4605-4d4f-927f-d1eafe06956f-ovsdbserver-nb" (OuterVolumeSpecName: "ovsdbserver-nb") pod "2a0b5b97-4605-4d4f-927f-d1eafe06956f" (UID: "2a0b5b97-4605-4d4f-927f-d1eafe06956f"). InnerVolumeSpecName "ovsdbserver-nb". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.754974 4967 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-nb\" (UniqueName: \"kubernetes.io/configmap/2a0b5b97-4605-4d4f-927f-d1eafe06956f-ovsdbserver-nb\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.755017 4967 reconciler_common.go:293] "Volume detached for volume \"dns-swift-storage-0\" (UniqueName: \"kubernetes.io/configmap/2a0b5b97-4605-4d4f-927f-d1eafe06956f-dns-swift-storage-0\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.755030 4967 reconciler_common.go:293] "Volume detached for volume \"ovsdbserver-sb\" (UniqueName: \"kubernetes.io/configmap/2a0b5b97-4605-4d4f-927f-d1eafe06956f-ovsdbserver-sb\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.755041 4967 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2a0b5b97-4605-4d4f-927f-d1eafe06956f-config\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.755054 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4rgrq\" (UniqueName: \"kubernetes.io/projected/2a0b5b97-4605-4d4f-927f-d1eafe06956f-kube-api-access-4rgrq\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.763354 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2a0b5b97-4605-4d4f-927f-d1eafe06956f-dns-svc" (OuterVolumeSpecName: "dns-svc") pod "2a0b5b97-4605-4d4f-927f-d1eafe06956f" (UID: "2a0b5b97-4605-4d4f-927f-d1eafe06956f"). InnerVolumeSpecName "dns-svc". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.770605 4967 scope.go:117] "RemoveContainer" containerID="050e060ac770a18c635b538f899e5289cf36e4d693568ba33a0b57bd48d0eb85" Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.803194 4967 scope.go:117] "RemoveContainer" containerID="e42a65cb81472bcfcb031eb8b1298f5038365ed72c88b70eac57614ce427dde4" Oct 11 04:10:36 crc kubenswrapper[4967]: E1011 04:10:36.803651 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e42a65cb81472bcfcb031eb8b1298f5038365ed72c88b70eac57614ce427dde4\": container with ID starting with e42a65cb81472bcfcb031eb8b1298f5038365ed72c88b70eac57614ce427dde4 not found: ID does not exist" containerID="e42a65cb81472bcfcb031eb8b1298f5038365ed72c88b70eac57614ce427dde4" Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.803687 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e42a65cb81472bcfcb031eb8b1298f5038365ed72c88b70eac57614ce427dde4"} err="failed to get container status \"e42a65cb81472bcfcb031eb8b1298f5038365ed72c88b70eac57614ce427dde4\": rpc error: code = NotFound desc = could not find container \"e42a65cb81472bcfcb031eb8b1298f5038365ed72c88b70eac57614ce427dde4\": container with ID starting with e42a65cb81472bcfcb031eb8b1298f5038365ed72c88b70eac57614ce427dde4 not found: ID does not exist" Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.803734 4967 scope.go:117] "RemoveContainer" containerID="050e060ac770a18c635b538f899e5289cf36e4d693568ba33a0b57bd48d0eb85" Oct 11 04:10:36 crc kubenswrapper[4967]: E1011 04:10:36.803999 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"050e060ac770a18c635b538f899e5289cf36e4d693568ba33a0b57bd48d0eb85\": container with ID starting with 050e060ac770a18c635b538f899e5289cf36e4d693568ba33a0b57bd48d0eb85 not found: ID does not exist" containerID="050e060ac770a18c635b538f899e5289cf36e4d693568ba33a0b57bd48d0eb85" Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.804032 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"050e060ac770a18c635b538f899e5289cf36e4d693568ba33a0b57bd48d0eb85"} err="failed to get container status \"050e060ac770a18c635b538f899e5289cf36e4d693568ba33a0b57bd48d0eb85\": rpc error: code = NotFound desc = could not find container \"050e060ac770a18c635b538f899e5289cf36e4d693568ba33a0b57bd48d0eb85\": container with ID starting with 050e060ac770a18c635b538f899e5289cf36e4d693568ba33a0b57bd48d0eb85 not found: ID does not exist" Oct 11 04:10:36 crc kubenswrapper[4967]: I1011 04:10:36.856822 4967 reconciler_common.go:293] "Volume detached for volume \"dns-svc\" (UniqueName: \"kubernetes.io/configmap/2a0b5b97-4605-4d4f-927f-d1eafe06956f-dns-svc\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:37 crc kubenswrapper[4967]: I1011 04:10:37.006635 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:10:37 crc kubenswrapper[4967]: I1011 04:10:37.048474 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-xbm92"] Oct 11 04:10:37 crc kubenswrapper[4967]: I1011 04:10:37.079410 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/dnsmasq-dns-bccf8f775-xbm92"] Oct 11 04:10:37 crc kubenswrapper[4967]: I1011 04:10:37.733063 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c47c6348-b68f-410e-93d7-c10eed165a34","Type":"ContainerStarted","Data":"45e134fd4051fb55271709362e381aadbe5947e570b2cfc03cfcf87270cc10ae"} Oct 11 04:10:38 crc kubenswrapper[4967]: I1011 04:10:38.837029 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a0b5b97-4605-4d4f-927f-d1eafe06956f" path="/var/lib/kubelet/pods/2a0b5b97-4605-4d4f-927f-d1eafe06956f/volumes" Oct 11 04:10:39 crc kubenswrapper[4967]: I1011 04:10:39.078764 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 11 04:10:39 crc kubenswrapper[4967]: I1011 04:10:39.763860 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c47c6348-b68f-410e-93d7-c10eed165a34","Type":"ContainerStarted","Data":"7099750e002d39b4925d2605d982b7ed35f860552bca64bd99f2e75ff215158d"} Oct 11 04:10:41 crc kubenswrapper[4967]: I1011 04:10:41.787971 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ceilometer-0" event={"ID":"c47c6348-b68f-410e-93d7-c10eed165a34","Type":"ContainerStarted","Data":"9df9b292664761bdd47793cef60f0f4e1b26f7dc3e1100a0bd5c0c19915148d6"} Oct 11 04:10:41 crc kubenswrapper[4967]: I1011 04:10:41.788488 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ceilometer-0" Oct 11 04:10:41 crc kubenswrapper[4967]: I1011 04:10:41.820840 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ceilometer-0" podStartSLOduration=3.632039971 podStartE2EDuration="8.820819921s" podCreationTimestamp="2025-10-11 04:10:33 +0000 UTC" firstStartedPulling="2025-10-11 04:10:35.551269133 +0000 UTC m=+1163.514478066" lastFinishedPulling="2025-10-11 04:10:40.740049073 +0000 UTC m=+1168.703258016" observedRunningTime="2025-10-11 04:10:41.809771455 +0000 UTC m=+1169.772980398" watchObservedRunningTime="2025-10-11 04:10:41.820819921 +0000 UTC m=+1169.784028854" Oct 11 04:10:42 crc kubenswrapper[4967]: I1011 04:10:42.007269 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:10:42 crc kubenswrapper[4967]: I1011 04:10:42.057011 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:10:42 crc kubenswrapper[4967]: I1011 04:10:42.838552 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-cell1-novncproxy-0" Oct 11 04:10:43 crc kubenswrapper[4967]: I1011 04:10:43.041353 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-cell1-cell-mapping-2nl8n"] Oct 11 04:10:43 crc kubenswrapper[4967]: E1011 04:10:43.042317 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a0b5b97-4605-4d4f-927f-d1eafe06956f" containerName="dnsmasq-dns" Oct 11 04:10:43 crc kubenswrapper[4967]: I1011 04:10:43.042344 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a0b5b97-4605-4d4f-927f-d1eafe06956f" containerName="dnsmasq-dns" Oct 11 04:10:43 crc kubenswrapper[4967]: E1011 04:10:43.042384 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a0b5b97-4605-4d4f-927f-d1eafe06956f" containerName="init" Oct 11 04:10:43 crc kubenswrapper[4967]: I1011 04:10:43.042394 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a0b5b97-4605-4d4f-927f-d1eafe06956f" containerName="init" Oct 11 04:10:43 crc kubenswrapper[4967]: I1011 04:10:43.042930 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a0b5b97-4605-4d4f-927f-d1eafe06956f" containerName="dnsmasq-dns" Oct 11 04:10:43 crc kubenswrapper[4967]: I1011 04:10:43.043994 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2nl8n" Oct 11 04:10:43 crc kubenswrapper[4967]: I1011 04:10:43.057419 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-scripts" Oct 11 04:10:43 crc kubenswrapper[4967]: I1011 04:10:43.057385 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-cell1-manage-config-data" Oct 11 04:10:43 crc kubenswrapper[4967]: I1011 04:10:43.076041 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-2nl8n"] Oct 11 04:10:43 crc kubenswrapper[4967]: I1011 04:10:43.177288 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11ef0366-73db-43a2-88c4-780d2a0084c9-config-data\") pod \"nova-cell1-cell-mapping-2nl8n\" (UID: \"11ef0366-73db-43a2-88c4-780d2a0084c9\") " pod="openstack/nova-cell1-cell-mapping-2nl8n" Oct 11 04:10:43 crc kubenswrapper[4967]: I1011 04:10:43.177357 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11ef0366-73db-43a2-88c4-780d2a0084c9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-2nl8n\" (UID: \"11ef0366-73db-43a2-88c4-780d2a0084c9\") " pod="openstack/nova-cell1-cell-mapping-2nl8n" Oct 11 04:10:43 crc kubenswrapper[4967]: I1011 04:10:43.177748 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gcq44\" (UniqueName: \"kubernetes.io/projected/11ef0366-73db-43a2-88c4-780d2a0084c9-kube-api-access-gcq44\") pod \"nova-cell1-cell-mapping-2nl8n\" (UID: \"11ef0366-73db-43a2-88c4-780d2a0084c9\") " pod="openstack/nova-cell1-cell-mapping-2nl8n" Oct 11 04:10:43 crc kubenswrapper[4967]: I1011 04:10:43.177865 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11ef0366-73db-43a2-88c4-780d2a0084c9-scripts\") pod \"nova-cell1-cell-mapping-2nl8n\" (UID: \"11ef0366-73db-43a2-88c4-780d2a0084c9\") " pod="openstack/nova-cell1-cell-mapping-2nl8n" Oct 11 04:10:43 crc kubenswrapper[4967]: I1011 04:10:43.279299 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gcq44\" (UniqueName: \"kubernetes.io/projected/11ef0366-73db-43a2-88c4-780d2a0084c9-kube-api-access-gcq44\") pod \"nova-cell1-cell-mapping-2nl8n\" (UID: \"11ef0366-73db-43a2-88c4-780d2a0084c9\") " pod="openstack/nova-cell1-cell-mapping-2nl8n" Oct 11 04:10:43 crc kubenswrapper[4967]: I1011 04:10:43.279692 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11ef0366-73db-43a2-88c4-780d2a0084c9-scripts\") pod \"nova-cell1-cell-mapping-2nl8n\" (UID: \"11ef0366-73db-43a2-88c4-780d2a0084c9\") " pod="openstack/nova-cell1-cell-mapping-2nl8n" Oct 11 04:10:43 crc kubenswrapper[4967]: I1011 04:10:43.279767 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11ef0366-73db-43a2-88c4-780d2a0084c9-config-data\") pod \"nova-cell1-cell-mapping-2nl8n\" (UID: \"11ef0366-73db-43a2-88c4-780d2a0084c9\") " pod="openstack/nova-cell1-cell-mapping-2nl8n" Oct 11 04:10:43 crc kubenswrapper[4967]: I1011 04:10:43.279814 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11ef0366-73db-43a2-88c4-780d2a0084c9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-2nl8n\" (UID: \"11ef0366-73db-43a2-88c4-780d2a0084c9\") " pod="openstack/nova-cell1-cell-mapping-2nl8n" Oct 11 04:10:43 crc kubenswrapper[4967]: I1011 04:10:43.285682 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11ef0366-73db-43a2-88c4-780d2a0084c9-scripts\") pod \"nova-cell1-cell-mapping-2nl8n\" (UID: \"11ef0366-73db-43a2-88c4-780d2a0084c9\") " pod="openstack/nova-cell1-cell-mapping-2nl8n" Oct 11 04:10:43 crc kubenswrapper[4967]: I1011 04:10:43.298286 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11ef0366-73db-43a2-88c4-780d2a0084c9-combined-ca-bundle\") pod \"nova-cell1-cell-mapping-2nl8n\" (UID: \"11ef0366-73db-43a2-88c4-780d2a0084c9\") " pod="openstack/nova-cell1-cell-mapping-2nl8n" Oct 11 04:10:43 crc kubenswrapper[4967]: I1011 04:10:43.303185 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11ef0366-73db-43a2-88c4-780d2a0084c9-config-data\") pod \"nova-cell1-cell-mapping-2nl8n\" (UID: \"11ef0366-73db-43a2-88c4-780d2a0084c9\") " pod="openstack/nova-cell1-cell-mapping-2nl8n" Oct 11 04:10:43 crc kubenswrapper[4967]: I1011 04:10:43.307178 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gcq44\" (UniqueName: \"kubernetes.io/projected/11ef0366-73db-43a2-88c4-780d2a0084c9-kube-api-access-gcq44\") pod \"nova-cell1-cell-mapping-2nl8n\" (UID: \"11ef0366-73db-43a2-88c4-780d2a0084c9\") " pod="openstack/nova-cell1-cell-mapping-2nl8n" Oct 11 04:10:43 crc kubenswrapper[4967]: I1011 04:10:43.379756 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2nl8n" Oct 11 04:10:43 crc kubenswrapper[4967]: I1011 04:10:43.955964 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-cell1-cell-mapping-2nl8n"] Oct 11 04:10:44 crc kubenswrapper[4967]: I1011 04:10:44.078706 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 11 04:10:44 crc kubenswrapper[4967]: I1011 04:10:44.119173 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 11 04:10:44 crc kubenswrapper[4967]: I1011 04:10:44.308713 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 11 04:10:44 crc kubenswrapper[4967]: I1011 04:10:44.308772 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 11 04:10:44 crc kubenswrapper[4967]: I1011 04:10:44.841990 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2nl8n" event={"ID":"11ef0366-73db-43a2-88c4-780d2a0084c9","Type":"ContainerStarted","Data":"910da87b7c970cfcc4b89b941d30c32a87f174c3e567e6f92be4f17019eb4749"} Oct 11 04:10:44 crc kubenswrapper[4967]: I1011 04:10:44.844295 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2nl8n" event={"ID":"11ef0366-73db-43a2-88c4-780d2a0084c9","Type":"ContainerStarted","Data":"eae93dfd4e6084b1fefbd16a0773f143d81d002e5f87033d36587e7b69e3b5ae"} Oct 11 04:10:44 crc kubenswrapper[4967]: I1011 04:10:44.860152 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 11 04:10:44 crc kubenswrapper[4967]: I1011 04:10:44.893315 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-cell1-cell-mapping-2nl8n" podStartSLOduration=1.893296343 podStartE2EDuration="1.893296343s" podCreationTimestamp="2025-10-11 04:10:43 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:10:44.83661139 +0000 UTC m=+1172.799820353" watchObservedRunningTime="2025-10-11 04:10:44.893296343 +0000 UTC m=+1172.856505276" Oct 11 04:10:45 crc kubenswrapper[4967]: I1011 04:10:45.325203 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.208:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 11 04:10:45 crc kubenswrapper[4967]: I1011 04:10:45.325203 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.208:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 11 04:10:48 crc kubenswrapper[4967]: I1011 04:10:48.886983 4967 generic.go:334] "Generic (PLEG): container finished" podID="11ef0366-73db-43a2-88c4-780d2a0084c9" containerID="910da87b7c970cfcc4b89b941d30c32a87f174c3e567e6f92be4f17019eb4749" exitCode=0 Oct 11 04:10:48 crc kubenswrapper[4967]: I1011 04:10:48.887426 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2nl8n" event={"ID":"11ef0366-73db-43a2-88c4-780d2a0084c9","Type":"ContainerDied","Data":"910da87b7c970cfcc4b89b941d30c32a87f174c3e567e6f92be4f17019eb4749"} Oct 11 04:10:50 crc kubenswrapper[4967]: I1011 04:10:50.273597 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2nl8n" Oct 11 04:10:50 crc kubenswrapper[4967]: I1011 04:10:50.320471 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gcq44\" (UniqueName: \"kubernetes.io/projected/11ef0366-73db-43a2-88c4-780d2a0084c9-kube-api-access-gcq44\") pod \"11ef0366-73db-43a2-88c4-780d2a0084c9\" (UID: \"11ef0366-73db-43a2-88c4-780d2a0084c9\") " Oct 11 04:10:50 crc kubenswrapper[4967]: I1011 04:10:50.320529 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11ef0366-73db-43a2-88c4-780d2a0084c9-combined-ca-bundle\") pod \"11ef0366-73db-43a2-88c4-780d2a0084c9\" (UID: \"11ef0366-73db-43a2-88c4-780d2a0084c9\") " Oct 11 04:10:50 crc kubenswrapper[4967]: I1011 04:10:50.320622 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11ef0366-73db-43a2-88c4-780d2a0084c9-scripts\") pod \"11ef0366-73db-43a2-88c4-780d2a0084c9\" (UID: \"11ef0366-73db-43a2-88c4-780d2a0084c9\") " Oct 11 04:10:50 crc kubenswrapper[4967]: I1011 04:10:50.320645 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11ef0366-73db-43a2-88c4-780d2a0084c9-config-data\") pod \"11ef0366-73db-43a2-88c4-780d2a0084c9\" (UID: \"11ef0366-73db-43a2-88c4-780d2a0084c9\") " Oct 11 04:10:50 crc kubenswrapper[4967]: I1011 04:10:50.326466 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11ef0366-73db-43a2-88c4-780d2a0084c9-scripts" (OuterVolumeSpecName: "scripts") pod "11ef0366-73db-43a2-88c4-780d2a0084c9" (UID: "11ef0366-73db-43a2-88c4-780d2a0084c9"). InnerVolumeSpecName "scripts". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:50 crc kubenswrapper[4967]: I1011 04:10:50.330348 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/11ef0366-73db-43a2-88c4-780d2a0084c9-kube-api-access-gcq44" (OuterVolumeSpecName: "kube-api-access-gcq44") pod "11ef0366-73db-43a2-88c4-780d2a0084c9" (UID: "11ef0366-73db-43a2-88c4-780d2a0084c9"). InnerVolumeSpecName "kube-api-access-gcq44". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:10:50 crc kubenswrapper[4967]: I1011 04:10:50.356523 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11ef0366-73db-43a2-88c4-780d2a0084c9-config-data" (OuterVolumeSpecName: "config-data") pod "11ef0366-73db-43a2-88c4-780d2a0084c9" (UID: "11ef0366-73db-43a2-88c4-780d2a0084c9"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:50 crc kubenswrapper[4967]: I1011 04:10:50.362410 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/11ef0366-73db-43a2-88c4-780d2a0084c9-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "11ef0366-73db-43a2-88c4-780d2a0084c9" (UID: "11ef0366-73db-43a2-88c4-780d2a0084c9"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:50 crc kubenswrapper[4967]: I1011 04:10:50.422435 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gcq44\" (UniqueName: \"kubernetes.io/projected/11ef0366-73db-43a2-88c4-780d2a0084c9-kube-api-access-gcq44\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:50 crc kubenswrapper[4967]: I1011 04:10:50.422469 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/11ef0366-73db-43a2-88c4-780d2a0084c9-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:50 crc kubenswrapper[4967]: I1011 04:10:50.422478 4967 reconciler_common.go:293] "Volume detached for volume \"scripts\" (UniqueName: \"kubernetes.io/secret/11ef0366-73db-43a2-88c4-780d2a0084c9-scripts\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:50 crc kubenswrapper[4967]: I1011 04:10:50.422499 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/11ef0366-73db-43a2-88c4-780d2a0084c9-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:50 crc kubenswrapper[4967]: I1011 04:10:50.906906 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-cell1-cell-mapping-2nl8n" event={"ID":"11ef0366-73db-43a2-88c4-780d2a0084c9","Type":"ContainerDied","Data":"eae93dfd4e6084b1fefbd16a0773f143d81d002e5f87033d36587e7b69e3b5ae"} Oct 11 04:10:50 crc kubenswrapper[4967]: I1011 04:10:50.907217 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="eae93dfd4e6084b1fefbd16a0773f143d81d002e5f87033d36587e7b69e3b5ae" Oct 11 04:10:50 crc kubenswrapper[4967]: I1011 04:10:50.907181 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-cell1-cell-mapping-2nl8n" Oct 11 04:10:51 crc kubenswrapper[4967]: I1011 04:10:51.121802 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 11 04:10:51 crc kubenswrapper[4967]: I1011 04:10:51.122053 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7" containerName="nova-api-log" containerID="cri-o://6322bd913c3c8da7f4f00e75fa8002b5ceda98f094f842d81283ab3ae5c2f7b0" gracePeriod=30 Oct 11 04:10:51 crc kubenswrapper[4967]: I1011 04:10:51.122169 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-api-0" podUID="64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7" containerName="nova-api-api" containerID="cri-o://0656e8fb8f6f7b37d2f167a44285df4d9003297f85025352040830189eaa3c1e" gracePeriod=30 Oct 11 04:10:51 crc kubenswrapper[4967]: I1011 04:10:51.134314 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 11 04:10:51 crc kubenswrapper[4967]: I1011 04:10:51.134522 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-scheduler-0" podUID="ecf46e75-f69c-4049-ada3-f74914ae8346" containerName="nova-scheduler-scheduler" containerID="cri-o://0ff59e5bf52f194cbaeaa5baec2131cf19c3f2cb5a5617318169fb63f10618d9" gracePeriod=30 Oct 11 04:10:51 crc kubenswrapper[4967]: I1011 04:10:51.183733 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 11 04:10:51 crc kubenswrapper[4967]: I1011 04:10:51.184040 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ebd723b4-b11e-46f6-a034-80269be7c26b" containerName="nova-metadata-log" containerID="cri-o://ffe47789cdc470ff6ccfced73645ce45dc1e8c178037e40d42b7385ff255a77b" gracePeriod=30 Oct 11 04:10:51 crc kubenswrapper[4967]: I1011 04:10:51.184131 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/nova-metadata-0" podUID="ebd723b4-b11e-46f6-a034-80269be7c26b" containerName="nova-metadata-metadata" containerID="cri-o://eb9488bfd76965f1d7ccd7edd22ded7df84fe1b36de84d5b0693c3cae286ae2b" gracePeriod=30 Oct 11 04:10:51 crc kubenswrapper[4967]: I1011 04:10:51.928167 4967 generic.go:334] "Generic (PLEG): container finished" podID="ebd723b4-b11e-46f6-a034-80269be7c26b" containerID="ffe47789cdc470ff6ccfced73645ce45dc1e8c178037e40d42b7385ff255a77b" exitCode=143 Oct 11 04:10:51 crc kubenswrapper[4967]: I1011 04:10:51.928245 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ebd723b4-b11e-46f6-a034-80269be7c26b","Type":"ContainerDied","Data":"ffe47789cdc470ff6ccfced73645ce45dc1e8c178037e40d42b7385ff255a77b"} Oct 11 04:10:51 crc kubenswrapper[4967]: I1011 04:10:51.930822 4967 generic.go:334] "Generic (PLEG): container finished" podID="64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7" containerID="6322bd913c3c8da7f4f00e75fa8002b5ceda98f094f842d81283ab3ae5c2f7b0" exitCode=143 Oct 11 04:10:51 crc kubenswrapper[4967]: I1011 04:10:51.930900 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7","Type":"ContainerDied","Data":"6322bd913c3c8da7f4f00e75fa8002b5ceda98f094f842d81283ab3ae5c2f7b0"} Oct 11 04:10:51 crc kubenswrapper[4967]: I1011 04:10:51.932356 4967 generic.go:334] "Generic (PLEG): container finished" podID="ecf46e75-f69c-4049-ada3-f74914ae8346" containerID="0ff59e5bf52f194cbaeaa5baec2131cf19c3f2cb5a5617318169fb63f10618d9" exitCode=0 Oct 11 04:10:51 crc kubenswrapper[4967]: I1011 04:10:51.932388 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ecf46e75-f69c-4049-ada3-f74914ae8346","Type":"ContainerDied","Data":"0ff59e5bf52f194cbaeaa5baec2131cf19c3f2cb5a5617318169fb63f10618d9"} Oct 11 04:10:52 crc kubenswrapper[4967]: I1011 04:10:52.269913 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 11 04:10:52 crc kubenswrapper[4967]: I1011 04:10:52.371199 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecf46e75-f69c-4049-ada3-f74914ae8346-config-data\") pod \"ecf46e75-f69c-4049-ada3-f74914ae8346\" (UID: \"ecf46e75-f69c-4049-ada3-f74914ae8346\") " Oct 11 04:10:52 crc kubenswrapper[4967]: I1011 04:10:52.371272 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fsqmd\" (UniqueName: \"kubernetes.io/projected/ecf46e75-f69c-4049-ada3-f74914ae8346-kube-api-access-fsqmd\") pod \"ecf46e75-f69c-4049-ada3-f74914ae8346\" (UID: \"ecf46e75-f69c-4049-ada3-f74914ae8346\") " Oct 11 04:10:52 crc kubenswrapper[4967]: I1011 04:10:52.371344 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecf46e75-f69c-4049-ada3-f74914ae8346-combined-ca-bundle\") pod \"ecf46e75-f69c-4049-ada3-f74914ae8346\" (UID: \"ecf46e75-f69c-4049-ada3-f74914ae8346\") " Oct 11 04:10:52 crc kubenswrapper[4967]: I1011 04:10:52.377714 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ecf46e75-f69c-4049-ada3-f74914ae8346-kube-api-access-fsqmd" (OuterVolumeSpecName: "kube-api-access-fsqmd") pod "ecf46e75-f69c-4049-ada3-f74914ae8346" (UID: "ecf46e75-f69c-4049-ada3-f74914ae8346"). InnerVolumeSpecName "kube-api-access-fsqmd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:10:52 crc kubenswrapper[4967]: I1011 04:10:52.406381 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecf46e75-f69c-4049-ada3-f74914ae8346-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ecf46e75-f69c-4049-ada3-f74914ae8346" (UID: "ecf46e75-f69c-4049-ada3-f74914ae8346"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:52 crc kubenswrapper[4967]: I1011 04:10:52.407782 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ecf46e75-f69c-4049-ada3-f74914ae8346-config-data" (OuterVolumeSpecName: "config-data") pod "ecf46e75-f69c-4049-ada3-f74914ae8346" (UID: "ecf46e75-f69c-4049-ada3-f74914ae8346"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:52 crc kubenswrapper[4967]: I1011 04:10:52.474354 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ecf46e75-f69c-4049-ada3-f74914ae8346-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:52 crc kubenswrapper[4967]: I1011 04:10:52.474397 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fsqmd\" (UniqueName: \"kubernetes.io/projected/ecf46e75-f69c-4049-ada3-f74914ae8346-kube-api-access-fsqmd\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:52 crc kubenswrapper[4967]: I1011 04:10:52.474408 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ecf46e75-f69c-4049-ada3-f74914ae8346-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:52 crc kubenswrapper[4967]: I1011 04:10:52.948008 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"ecf46e75-f69c-4049-ada3-f74914ae8346","Type":"ContainerDied","Data":"3dfc40a7aa0ede18b4dd70eefc3e4f423c75904366af6b65d8c96ade852176ff"} Oct 11 04:10:52 crc kubenswrapper[4967]: I1011 04:10:52.948059 4967 scope.go:117] "RemoveContainer" containerID="0ff59e5bf52f194cbaeaa5baec2131cf19c3f2cb5a5617318169fb63f10618d9" Oct 11 04:10:52 crc kubenswrapper[4967]: I1011 04:10:52.948195 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 11 04:10:52 crc kubenswrapper[4967]: I1011 04:10:52.979996 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-scheduler-0"] Oct 11 04:10:52 crc kubenswrapper[4967]: I1011 04:10:52.990186 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-scheduler-0"] Oct 11 04:10:53 crc kubenswrapper[4967]: I1011 04:10:53.007180 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-scheduler-0"] Oct 11 04:10:53 crc kubenswrapper[4967]: E1011 04:10:53.007656 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="11ef0366-73db-43a2-88c4-780d2a0084c9" containerName="nova-manage" Oct 11 04:10:53 crc kubenswrapper[4967]: I1011 04:10:53.007718 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="11ef0366-73db-43a2-88c4-780d2a0084c9" containerName="nova-manage" Oct 11 04:10:53 crc kubenswrapper[4967]: E1011 04:10:53.007778 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ecf46e75-f69c-4049-ada3-f74914ae8346" containerName="nova-scheduler-scheduler" Oct 11 04:10:53 crc kubenswrapper[4967]: I1011 04:10:53.007822 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="ecf46e75-f69c-4049-ada3-f74914ae8346" containerName="nova-scheduler-scheduler" Oct 11 04:10:53 crc kubenswrapper[4967]: I1011 04:10:53.008086 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="11ef0366-73db-43a2-88c4-780d2a0084c9" containerName="nova-manage" Oct 11 04:10:53 crc kubenswrapper[4967]: I1011 04:10:53.008164 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="ecf46e75-f69c-4049-ada3-f74914ae8346" containerName="nova-scheduler-scheduler" Oct 11 04:10:53 crc kubenswrapper[4967]: I1011 04:10:53.008876 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 11 04:10:53 crc kubenswrapper[4967]: I1011 04:10:53.011386 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-scheduler-config-data" Oct 11 04:10:53 crc kubenswrapper[4967]: I1011 04:10:53.029112 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 11 04:10:53 crc kubenswrapper[4967]: I1011 04:10:53.084190 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/388b5da4-c3cd-43f3-b839-cfd0b6c2a8f8-config-data\") pod \"nova-scheduler-0\" (UID: \"388b5da4-c3cd-43f3-b839-cfd0b6c2a8f8\") " pod="openstack/nova-scheduler-0" Oct 11 04:10:53 crc kubenswrapper[4967]: I1011 04:10:53.084236 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z865d\" (UniqueName: \"kubernetes.io/projected/388b5da4-c3cd-43f3-b839-cfd0b6c2a8f8-kube-api-access-z865d\") pod \"nova-scheduler-0\" (UID: \"388b5da4-c3cd-43f3-b839-cfd0b6c2a8f8\") " pod="openstack/nova-scheduler-0" Oct 11 04:10:53 crc kubenswrapper[4967]: I1011 04:10:53.084356 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/388b5da4-c3cd-43f3-b839-cfd0b6c2a8f8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"388b5da4-c3cd-43f3-b839-cfd0b6c2a8f8\") " pod="openstack/nova-scheduler-0" Oct 11 04:10:53 crc kubenswrapper[4967]: I1011 04:10:53.186464 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/388b5da4-c3cd-43f3-b839-cfd0b6c2a8f8-config-data\") pod \"nova-scheduler-0\" (UID: \"388b5da4-c3cd-43f3-b839-cfd0b6c2a8f8\") " pod="openstack/nova-scheduler-0" Oct 11 04:10:53 crc kubenswrapper[4967]: I1011 04:10:53.186570 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z865d\" (UniqueName: \"kubernetes.io/projected/388b5da4-c3cd-43f3-b839-cfd0b6c2a8f8-kube-api-access-z865d\") pod \"nova-scheduler-0\" (UID: \"388b5da4-c3cd-43f3-b839-cfd0b6c2a8f8\") " pod="openstack/nova-scheduler-0" Oct 11 04:10:53 crc kubenswrapper[4967]: I1011 04:10:53.187064 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/388b5da4-c3cd-43f3-b839-cfd0b6c2a8f8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"388b5da4-c3cd-43f3-b839-cfd0b6c2a8f8\") " pod="openstack/nova-scheduler-0" Oct 11 04:10:53 crc kubenswrapper[4967]: I1011 04:10:53.199178 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/388b5da4-c3cd-43f3-b839-cfd0b6c2a8f8-config-data\") pod \"nova-scheduler-0\" (UID: \"388b5da4-c3cd-43f3-b839-cfd0b6c2a8f8\") " pod="openstack/nova-scheduler-0" Oct 11 04:10:53 crc kubenswrapper[4967]: I1011 04:10:53.199183 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/388b5da4-c3cd-43f3-b839-cfd0b6c2a8f8-combined-ca-bundle\") pod \"nova-scheduler-0\" (UID: \"388b5da4-c3cd-43f3-b839-cfd0b6c2a8f8\") " pod="openstack/nova-scheduler-0" Oct 11 04:10:53 crc kubenswrapper[4967]: I1011 04:10:53.205005 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z865d\" (UniqueName: \"kubernetes.io/projected/388b5da4-c3cd-43f3-b839-cfd0b6c2a8f8-kube-api-access-z865d\") pod \"nova-scheduler-0\" (UID: \"388b5da4-c3cd-43f3-b839-cfd0b6c2a8f8\") " pod="openstack/nova-scheduler-0" Oct 11 04:10:53 crc kubenswrapper[4967]: I1011 04:10:53.333670 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-scheduler-0" Oct 11 04:10:53 crc kubenswrapper[4967]: I1011 04:10:53.786529 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-scheduler-0"] Oct 11 04:10:53 crc kubenswrapper[4967]: I1011 04:10:53.973486 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"388b5da4-c3cd-43f3-b839-cfd0b6c2a8f8","Type":"ContainerStarted","Data":"50061b50d82e9ac70acd5ccf89ac55e41a4d866735cc961247ddf35db68c8129"} Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.321139 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="ebd723b4-b11e-46f6-a034-80269be7c26b" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": read tcp 10.217.0.2:49032->10.217.0.202:8775: read: connection reset by peer" Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.321267 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/nova-metadata-0" podUID="ebd723b4-b11e-46f6-a034-80269be7c26b" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.202:8775/\": read tcp 10.217.0.2:49046->10.217.0.202:8775: read: connection reset by peer" Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.778748 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.785345 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.818066 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-internal-tls-certs\") pod \"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7\" (UID: \"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7\") " Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.818144 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-config-data\") pod \"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7\" (UID: \"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7\") " Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.818192 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rmfzr\" (UniqueName: \"kubernetes.io/projected/ebd723b4-b11e-46f6-a034-80269be7c26b-kube-api-access-rmfzr\") pod \"ebd723b4-b11e-46f6-a034-80269be7c26b\" (UID: \"ebd723b4-b11e-46f6-a034-80269be7c26b\") " Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.818217 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-combined-ca-bundle\") pod \"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7\" (UID: \"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7\") " Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.818254 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-logs\") pod \"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7\" (UID: \"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7\") " Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.818303 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebd723b4-b11e-46f6-a034-80269be7c26b-logs\") pod \"ebd723b4-b11e-46f6-a034-80269be7c26b\" (UID: \"ebd723b4-b11e-46f6-a034-80269be7c26b\") " Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.818466 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebd723b4-b11e-46f6-a034-80269be7c26b-nova-metadata-tls-certs\") pod \"ebd723b4-b11e-46f6-a034-80269be7c26b\" (UID: \"ebd723b4-b11e-46f6-a034-80269be7c26b\") " Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.818518 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebd723b4-b11e-46f6-a034-80269be7c26b-combined-ca-bundle\") pod \"ebd723b4-b11e-46f6-a034-80269be7c26b\" (UID: \"ebd723b4-b11e-46f6-a034-80269be7c26b\") " Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.818554 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebd723b4-b11e-46f6-a034-80269be7c26b-config-data\") pod \"ebd723b4-b11e-46f6-a034-80269be7c26b\" (UID: \"ebd723b4-b11e-46f6-a034-80269be7c26b\") " Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.818577 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-public-tls-certs\") pod \"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7\" (UID: \"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7\") " Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.818641 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jxkkv\" (UniqueName: \"kubernetes.io/projected/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-kube-api-access-jxkkv\") pod \"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7\" (UID: \"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7\") " Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.818824 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-logs" (OuterVolumeSpecName: "logs") pod "64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7" (UID: "64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.818911 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/ebd723b4-b11e-46f6-a034-80269be7c26b-logs" (OuterVolumeSpecName: "logs") pod "ebd723b4-b11e-46f6-a034-80269be7c26b" (UID: "ebd723b4-b11e-46f6-a034-80269be7c26b"). InnerVolumeSpecName "logs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.819172 4967 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/ebd723b4-b11e-46f6-a034-80269be7c26b-logs\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.819189 4967 reconciler_common.go:293] "Volume detached for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-logs\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.828123 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-kube-api-access-jxkkv" (OuterVolumeSpecName: "kube-api-access-jxkkv") pod "64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7" (UID: "64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7"). InnerVolumeSpecName "kube-api-access-jxkkv". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.828315 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/ebd723b4-b11e-46f6-a034-80269be7c26b-kube-api-access-rmfzr" (OuterVolumeSpecName: "kube-api-access-rmfzr") pod "ebd723b4-b11e-46f6-a034-80269be7c26b" (UID: "ebd723b4-b11e-46f6-a034-80269be7c26b"). InnerVolumeSpecName "kube-api-access-rmfzr". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.833013 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ecf46e75-f69c-4049-ada3-f74914ae8346" path="/var/lib/kubelet/pods/ecf46e75-f69c-4049-ada3-f74914ae8346/volumes" Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.873281 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebd723b4-b11e-46f6-a034-80269be7c26b-config-data" (OuterVolumeSpecName: "config-data") pod "ebd723b4-b11e-46f6-a034-80269be7c26b" (UID: "ebd723b4-b11e-46f6-a034-80269be7c26b"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.879657 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7" (UID: "64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.899834 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebd723b4-b11e-46f6-a034-80269be7c26b-combined-ca-bundle" (OuterVolumeSpecName: "combined-ca-bundle") pod "ebd723b4-b11e-46f6-a034-80269be7c26b" (UID: "ebd723b4-b11e-46f6-a034-80269be7c26b"). InnerVolumeSpecName "combined-ca-bundle". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.902828 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-config-data" (OuterVolumeSpecName: "config-data") pod "64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7" (UID: "64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.922490 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/ebd723b4-b11e-46f6-a034-80269be7c26b-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.922660 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/ebd723b4-b11e-46f6-a034-80269be7c26b-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.922713 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jxkkv\" (UniqueName: \"kubernetes.io/projected/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-kube-api-access-jxkkv\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.922763 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.922811 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rmfzr\" (UniqueName: \"kubernetes.io/projected/ebd723b4-b11e-46f6-a034-80269be7c26b-kube-api-access-rmfzr\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.922873 4967 reconciler_common.go:293] "Volume detached for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-combined-ca-bundle\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.924521 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-public-tls-certs" (OuterVolumeSpecName: "public-tls-certs") pod "64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7" (UID: "64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7"). InnerVolumeSpecName "public-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.930930 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-internal-tls-certs" (OuterVolumeSpecName: "internal-tls-certs") pod "64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7" (UID: "64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7"). InnerVolumeSpecName "internal-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.932014 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/ebd723b4-b11e-46f6-a034-80269be7c26b-nova-metadata-tls-certs" (OuterVolumeSpecName: "nova-metadata-tls-certs") pod "ebd723b4-b11e-46f6-a034-80269be7c26b" (UID: "ebd723b4-b11e-46f6-a034-80269be7c26b"). InnerVolumeSpecName "nova-metadata-tls-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.992979 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-scheduler-0" event={"ID":"388b5da4-c3cd-43f3-b839-cfd0b6c2a8f8","Type":"ContainerStarted","Data":"b1174d8f5a79563581a2ac56d53629b556d9701cadc0d5e09bd8672aee24f0b9"} Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.995853 4967 generic.go:334] "Generic (PLEG): container finished" podID="ebd723b4-b11e-46f6-a034-80269be7c26b" containerID="eb9488bfd76965f1d7ccd7edd22ded7df84fe1b36de84d5b0693c3cae286ae2b" exitCode=0 Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.995907 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ebd723b4-b11e-46f6-a034-80269be7c26b","Type":"ContainerDied","Data":"eb9488bfd76965f1d7ccd7edd22ded7df84fe1b36de84d5b0693c3cae286ae2b"} Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.995931 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"ebd723b4-b11e-46f6-a034-80269be7c26b","Type":"ContainerDied","Data":"19bbed4ff049637f8c993f16fc2b92da04c401a6a7d7c2912ee5992a218a7d17"} Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.995949 4967 scope.go:117] "RemoveContainer" containerID="eb9488bfd76965f1d7ccd7edd22ded7df84fe1b36de84d5b0693c3cae286ae2b" Oct 11 04:10:54 crc kubenswrapper[4967]: I1011 04:10:54.996085 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.009046 4967 generic.go:334] "Generic (PLEG): container finished" podID="64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7" containerID="0656e8fb8f6f7b37d2f167a44285df4d9003297f85025352040830189eaa3c1e" exitCode=0 Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.009111 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.009130 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7","Type":"ContainerDied","Data":"0656e8fb8f6f7b37d2f167a44285df4d9003297f85025352040830189eaa3c1e"} Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.009879 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7","Type":"ContainerDied","Data":"3dc831d1cb19848fa8ff46ed907560ccdc0c2a5a941baa673d1cc5242086fe98"} Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.018662 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-scheduler-0" podStartSLOduration=3.018646261 podStartE2EDuration="3.018646261s" podCreationTimestamp="2025-10-11 04:10:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:10:55.01638474 +0000 UTC m=+1182.979593673" watchObservedRunningTime="2025-10-11 04:10:55.018646261 +0000 UTC m=+1182.981855195" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.024699 4967 reconciler_common.go:293] "Volume detached for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-internal-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.024888 4967 reconciler_common.go:293] "Volume detached for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/ebd723b4-b11e-46f6-a034-80269be7c26b-nova-metadata-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.024971 4967 reconciler_common.go:293] "Volume detached for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7-public-tls-certs\") on node \"crc\" DevicePath \"\"" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.034847 4967 scope.go:117] "RemoveContainer" containerID="ffe47789cdc470ff6ccfced73645ce45dc1e8c178037e40d42b7385ff255a77b" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.047418 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-metadata-0"] Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.062979 4967 scope.go:117] "RemoveContainer" containerID="eb9488bfd76965f1d7ccd7edd22ded7df84fe1b36de84d5b0693c3cae286ae2b" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.063105 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-metadata-0"] Oct 11 04:10:55 crc kubenswrapper[4967]: E1011 04:10:55.064334 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"eb9488bfd76965f1d7ccd7edd22ded7df84fe1b36de84d5b0693c3cae286ae2b\": container with ID starting with eb9488bfd76965f1d7ccd7edd22ded7df84fe1b36de84d5b0693c3cae286ae2b not found: ID does not exist" containerID="eb9488bfd76965f1d7ccd7edd22ded7df84fe1b36de84d5b0693c3cae286ae2b" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.064361 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"eb9488bfd76965f1d7ccd7edd22ded7df84fe1b36de84d5b0693c3cae286ae2b"} err="failed to get container status \"eb9488bfd76965f1d7ccd7edd22ded7df84fe1b36de84d5b0693c3cae286ae2b\": rpc error: code = NotFound desc = could not find container \"eb9488bfd76965f1d7ccd7edd22ded7df84fe1b36de84d5b0693c3cae286ae2b\": container with ID starting with eb9488bfd76965f1d7ccd7edd22ded7df84fe1b36de84d5b0693c3cae286ae2b not found: ID does not exist" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.064380 4967 scope.go:117] "RemoveContainer" containerID="ffe47789cdc470ff6ccfced73645ce45dc1e8c178037e40d42b7385ff255a77b" Oct 11 04:10:55 crc kubenswrapper[4967]: E1011 04:10:55.068192 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ffe47789cdc470ff6ccfced73645ce45dc1e8c178037e40d42b7385ff255a77b\": container with ID starting with ffe47789cdc470ff6ccfced73645ce45dc1e8c178037e40d42b7385ff255a77b not found: ID does not exist" containerID="ffe47789cdc470ff6ccfced73645ce45dc1e8c178037e40d42b7385ff255a77b" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.068233 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ffe47789cdc470ff6ccfced73645ce45dc1e8c178037e40d42b7385ff255a77b"} err="failed to get container status \"ffe47789cdc470ff6ccfced73645ce45dc1e8c178037e40d42b7385ff255a77b\": rpc error: code = NotFound desc = could not find container \"ffe47789cdc470ff6ccfced73645ce45dc1e8c178037e40d42b7385ff255a77b\": container with ID starting with ffe47789cdc470ff6ccfced73645ce45dc1e8c178037e40d42b7385ff255a77b not found: ID does not exist" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.068257 4967 scope.go:117] "RemoveContainer" containerID="0656e8fb8f6f7b37d2f167a44285df4d9003297f85025352040830189eaa3c1e" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.075167 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-0"] Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.091326 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-metadata-0"] Oct 11 04:10:55 crc kubenswrapper[4967]: E1011 04:10:55.091706 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebd723b4-b11e-46f6-a034-80269be7c26b" containerName="nova-metadata-log" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.091724 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebd723b4-b11e-46f6-a034-80269be7c26b" containerName="nova-metadata-log" Oct 11 04:10:55 crc kubenswrapper[4967]: E1011 04:10:55.091744 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7" containerName="nova-api-api" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.091750 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7" containerName="nova-api-api" Oct 11 04:10:55 crc kubenswrapper[4967]: E1011 04:10:55.091762 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="ebd723b4-b11e-46f6-a034-80269be7c26b" containerName="nova-metadata-metadata" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.091769 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="ebd723b4-b11e-46f6-a034-80269be7c26b" containerName="nova-metadata-metadata" Oct 11 04:10:55 crc kubenswrapper[4967]: E1011 04:10:55.091796 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7" containerName="nova-api-log" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.091802 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7" containerName="nova-api-log" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.091976 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebd723b4-b11e-46f6-a034-80269be7c26b" containerName="nova-metadata-log" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.091991 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7" containerName="nova-api-api" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.092009 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="ebd723b4-b11e-46f6-a034-80269be7c26b" containerName="nova-metadata-metadata" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.092025 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7" containerName="nova-api-log" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.093010 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.095479 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-metadata-config-data" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.102253 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-0"] Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.117189 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.119661 4967 scope.go:117] "RemoveContainer" containerID="6322bd913c3c8da7f4f00e75fa8002b5ceda98f094f842d81283ab3ae5c2f7b0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.119905 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-metadata-internal-svc" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.126282 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e701cd5-1cf9-4ed0-87c6-2023c41e7982-config-data\") pod \"nova-metadata-0\" (UID: \"8e701cd5-1cf9-4ed0-87c6-2023c41e7982\") " pod="openstack/nova-metadata-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.126362 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8hlnv\" (UniqueName: \"kubernetes.io/projected/8e701cd5-1cf9-4ed0-87c6-2023c41e7982-kube-api-access-8hlnv\") pod \"nova-metadata-0\" (UID: \"8e701cd5-1cf9-4ed0-87c6-2023c41e7982\") " pod="openstack/nova-metadata-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.126381 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e701cd5-1cf9-4ed0-87c6-2023c41e7982-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8e701cd5-1cf9-4ed0-87c6-2023c41e7982\") " pod="openstack/nova-metadata-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.126399 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e701cd5-1cf9-4ed0-87c6-2023c41e7982-logs\") pod \"nova-metadata-0\" (UID: \"8e701cd5-1cf9-4ed0-87c6-2023c41e7982\") " pod="openstack/nova-metadata-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.126446 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e701cd5-1cf9-4ed0-87c6-2023c41e7982-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8e701cd5-1cf9-4ed0-87c6-2023c41e7982\") " pod="openstack/nova-metadata-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.126539 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/nova-api-0"] Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.127901 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.133553 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-internal-svc" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.134812 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-nova-public-svc" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.133873 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"nova-api-config-data" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.137056 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.160267 4967 scope.go:117] "RemoveContainer" containerID="0656e8fb8f6f7b37d2f167a44285df4d9003297f85025352040830189eaa3c1e" Oct 11 04:10:55 crc kubenswrapper[4967]: E1011 04:10:55.160775 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0656e8fb8f6f7b37d2f167a44285df4d9003297f85025352040830189eaa3c1e\": container with ID starting with 0656e8fb8f6f7b37d2f167a44285df4d9003297f85025352040830189eaa3c1e not found: ID does not exist" containerID="0656e8fb8f6f7b37d2f167a44285df4d9003297f85025352040830189eaa3c1e" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.160847 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0656e8fb8f6f7b37d2f167a44285df4d9003297f85025352040830189eaa3c1e"} err="failed to get container status \"0656e8fb8f6f7b37d2f167a44285df4d9003297f85025352040830189eaa3c1e\": rpc error: code = NotFound desc = could not find container \"0656e8fb8f6f7b37d2f167a44285df4d9003297f85025352040830189eaa3c1e\": container with ID starting with 0656e8fb8f6f7b37d2f167a44285df4d9003297f85025352040830189eaa3c1e not found: ID does not exist" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.160888 4967 scope.go:117] "RemoveContainer" containerID="6322bd913c3c8da7f4f00e75fa8002b5ceda98f094f842d81283ab3ae5c2f7b0" Oct 11 04:10:55 crc kubenswrapper[4967]: E1011 04:10:55.161371 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6322bd913c3c8da7f4f00e75fa8002b5ceda98f094f842d81283ab3ae5c2f7b0\": container with ID starting with 6322bd913c3c8da7f4f00e75fa8002b5ceda98f094f842d81283ab3ae5c2f7b0 not found: ID does not exist" containerID="6322bd913c3c8da7f4f00e75fa8002b5ceda98f094f842d81283ab3ae5c2f7b0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.161492 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6322bd913c3c8da7f4f00e75fa8002b5ceda98f094f842d81283ab3ae5c2f7b0"} err="failed to get container status \"6322bd913c3c8da7f4f00e75fa8002b5ceda98f094f842d81283ab3ae5c2f7b0\": rpc error: code = NotFound desc = could not find container \"6322bd913c3c8da7f4f00e75fa8002b5ceda98f094f842d81283ab3ae5c2f7b0\": container with ID starting with 6322bd913c3c8da7f4f00e75fa8002b5ceda98f094f842d81283ab3ae5c2f7b0 not found: ID does not exist" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.228208 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c3cc259-3802-4172-8ae6-446cfbcd4ca5-config-data\") pod \"nova-api-0\" (UID: \"4c3cc259-3802-4172-8ae6-446cfbcd4ca5\") " pod="openstack/nova-api-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.228318 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c3cc259-3802-4172-8ae6-446cfbcd4ca5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4c3cc259-3802-4172-8ae6-446cfbcd4ca5\") " pod="openstack/nova-api-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.228357 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8hlnv\" (UniqueName: \"kubernetes.io/projected/8e701cd5-1cf9-4ed0-87c6-2023c41e7982-kube-api-access-8hlnv\") pod \"nova-metadata-0\" (UID: \"8e701cd5-1cf9-4ed0-87c6-2023c41e7982\") " pod="openstack/nova-metadata-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.228375 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e701cd5-1cf9-4ed0-87c6-2023c41e7982-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8e701cd5-1cf9-4ed0-87c6-2023c41e7982\") " pod="openstack/nova-metadata-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.228399 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e701cd5-1cf9-4ed0-87c6-2023c41e7982-logs\") pod \"nova-metadata-0\" (UID: \"8e701cd5-1cf9-4ed0-87c6-2023c41e7982\") " pod="openstack/nova-metadata-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.228444 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c3cc259-3802-4172-8ae6-446cfbcd4ca5-public-tls-certs\") pod \"nova-api-0\" (UID: \"4c3cc259-3802-4172-8ae6-446cfbcd4ca5\") " pod="openstack/nova-api-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.228467 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e701cd5-1cf9-4ed0-87c6-2023c41e7982-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8e701cd5-1cf9-4ed0-87c6-2023c41e7982\") " pod="openstack/nova-metadata-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.228499 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c3cc259-3802-4172-8ae6-446cfbcd4ca5-logs\") pod \"nova-api-0\" (UID: \"4c3cc259-3802-4172-8ae6-446cfbcd4ca5\") " pod="openstack/nova-api-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.228556 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e701cd5-1cf9-4ed0-87c6-2023c41e7982-config-data\") pod \"nova-metadata-0\" (UID: \"8e701cd5-1cf9-4ed0-87c6-2023c41e7982\") " pod="openstack/nova-metadata-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.228585 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4j4tk\" (UniqueName: \"kubernetes.io/projected/4c3cc259-3802-4172-8ae6-446cfbcd4ca5-kube-api-access-4j4tk\") pod \"nova-api-0\" (UID: \"4c3cc259-3802-4172-8ae6-446cfbcd4ca5\") " pod="openstack/nova-api-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.228616 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c3cc259-3802-4172-8ae6-446cfbcd4ca5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4c3cc259-3802-4172-8ae6-446cfbcd4ca5\") " pod="openstack/nova-api-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.229583 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/8e701cd5-1cf9-4ed0-87c6-2023c41e7982-logs\") pod \"nova-metadata-0\" (UID: \"8e701cd5-1cf9-4ed0-87c6-2023c41e7982\") " pod="openstack/nova-metadata-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.232989 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/8e701cd5-1cf9-4ed0-87c6-2023c41e7982-combined-ca-bundle\") pod \"nova-metadata-0\" (UID: \"8e701cd5-1cf9-4ed0-87c6-2023c41e7982\") " pod="openstack/nova-metadata-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.233354 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nova-metadata-tls-certs\" (UniqueName: \"kubernetes.io/secret/8e701cd5-1cf9-4ed0-87c6-2023c41e7982-nova-metadata-tls-certs\") pod \"nova-metadata-0\" (UID: \"8e701cd5-1cf9-4ed0-87c6-2023c41e7982\") " pod="openstack/nova-metadata-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.240061 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/8e701cd5-1cf9-4ed0-87c6-2023c41e7982-config-data\") pod \"nova-metadata-0\" (UID: \"8e701cd5-1cf9-4ed0-87c6-2023c41e7982\") " pod="openstack/nova-metadata-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.245124 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8hlnv\" (UniqueName: \"kubernetes.io/projected/8e701cd5-1cf9-4ed0-87c6-2023c41e7982-kube-api-access-8hlnv\") pod \"nova-metadata-0\" (UID: \"8e701cd5-1cf9-4ed0-87c6-2023c41e7982\") " pod="openstack/nova-metadata-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.330328 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c3cc259-3802-4172-8ae6-446cfbcd4ca5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4c3cc259-3802-4172-8ae6-446cfbcd4ca5\") " pod="openstack/nova-api-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.330830 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c3cc259-3802-4172-8ae6-446cfbcd4ca5-config-data\") pod \"nova-api-0\" (UID: \"4c3cc259-3802-4172-8ae6-446cfbcd4ca5\") " pod="openstack/nova-api-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.330868 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c3cc259-3802-4172-8ae6-446cfbcd4ca5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4c3cc259-3802-4172-8ae6-446cfbcd4ca5\") " pod="openstack/nova-api-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.330932 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c3cc259-3802-4172-8ae6-446cfbcd4ca5-public-tls-certs\") pod \"nova-api-0\" (UID: \"4c3cc259-3802-4172-8ae6-446cfbcd4ca5\") " pod="openstack/nova-api-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.330969 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c3cc259-3802-4172-8ae6-446cfbcd4ca5-logs\") pod \"nova-api-0\" (UID: \"4c3cc259-3802-4172-8ae6-446cfbcd4ca5\") " pod="openstack/nova-api-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.331050 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4j4tk\" (UniqueName: \"kubernetes.io/projected/4c3cc259-3802-4172-8ae6-446cfbcd4ca5-kube-api-access-4j4tk\") pod \"nova-api-0\" (UID: \"4c3cc259-3802-4172-8ae6-446cfbcd4ca5\") " pod="openstack/nova-api-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.331919 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"logs\" (UniqueName: \"kubernetes.io/empty-dir/4c3cc259-3802-4172-8ae6-446cfbcd4ca5-logs\") pod \"nova-api-0\" (UID: \"4c3cc259-3802-4172-8ae6-446cfbcd4ca5\") " pod="openstack/nova-api-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.333902 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"internal-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c3cc259-3802-4172-8ae6-446cfbcd4ca5-internal-tls-certs\") pod \"nova-api-0\" (UID: \"4c3cc259-3802-4172-8ae6-446cfbcd4ca5\") " pod="openstack/nova-api-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.334594 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"combined-ca-bundle\" (UniqueName: \"kubernetes.io/secret/4c3cc259-3802-4172-8ae6-446cfbcd4ca5-combined-ca-bundle\") pod \"nova-api-0\" (UID: \"4c3cc259-3802-4172-8ae6-446cfbcd4ca5\") " pod="openstack/nova-api-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.334894 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/secret/4c3cc259-3802-4172-8ae6-446cfbcd4ca5-config-data\") pod \"nova-api-0\" (UID: \"4c3cc259-3802-4172-8ae6-446cfbcd4ca5\") " pod="openstack/nova-api-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.335579 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"public-tls-certs\" (UniqueName: \"kubernetes.io/secret/4c3cc259-3802-4172-8ae6-446cfbcd4ca5-public-tls-certs\") pod \"nova-api-0\" (UID: \"4c3cc259-3802-4172-8ae6-446cfbcd4ca5\") " pod="openstack/nova-api-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.354609 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4j4tk\" (UniqueName: \"kubernetes.io/projected/4c3cc259-3802-4172-8ae6-446cfbcd4ca5-kube-api-access-4j4tk\") pod \"nova-api-0\" (UID: \"4c3cc259-3802-4172-8ae6-446cfbcd4ca5\") " pod="openstack/nova-api-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.421783 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-metadata-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.454459 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/nova-api-0" Oct 11 04:10:55 crc kubenswrapper[4967]: I1011 04:10:55.937189 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-metadata-0"] Oct 11 04:10:55 crc kubenswrapper[4967]: W1011 04:10:55.944530 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod8e701cd5_1cf9_4ed0_87c6_2023c41e7982.slice/crio-93cd83b3a5014ae1fe317bea34f5beed0fea37644fa1a703425e076ec94bc345 WatchSource:0}: Error finding container 93cd83b3a5014ae1fe317bea34f5beed0fea37644fa1a703425e076ec94bc345: Status 404 returned error can't find the container with id 93cd83b3a5014ae1fe317bea34f5beed0fea37644fa1a703425e076ec94bc345 Oct 11 04:10:56 crc kubenswrapper[4967]: I1011 04:10:56.006499 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/nova-api-0"] Oct 11 04:10:56 crc kubenswrapper[4967]: I1011 04:10:56.033196 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4c3cc259-3802-4172-8ae6-446cfbcd4ca5","Type":"ContainerStarted","Data":"b144d0d131a9d2b3c3ab74defc83726bd488cb48466fa13a53b652b2a17adc17"} Oct 11 04:10:56 crc kubenswrapper[4967]: I1011 04:10:56.044796 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8e701cd5-1cf9-4ed0-87c6-2023c41e7982","Type":"ContainerStarted","Data":"93cd83b3a5014ae1fe317bea34f5beed0fea37644fa1a703425e076ec94bc345"} Oct 11 04:10:56 crc kubenswrapper[4967]: I1011 04:10:56.829285 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7" path="/var/lib/kubelet/pods/64d9a7d9-e877-4f21-9d9f-dd9d9a6789f7/volumes" Oct 11 04:10:56 crc kubenswrapper[4967]: I1011 04:10:56.830426 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="ebd723b4-b11e-46f6-a034-80269be7c26b" path="/var/lib/kubelet/pods/ebd723b4-b11e-46f6-a034-80269be7c26b/volumes" Oct 11 04:10:57 crc kubenswrapper[4967]: I1011 04:10:57.058409 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8e701cd5-1cf9-4ed0-87c6-2023c41e7982","Type":"ContainerStarted","Data":"96334b3892fa791c4cbd7109056952c8c6d71ef4fc69b97884fdfca24df57b04"} Oct 11 04:10:57 crc kubenswrapper[4967]: I1011 04:10:57.058483 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-metadata-0" event={"ID":"8e701cd5-1cf9-4ed0-87c6-2023c41e7982","Type":"ContainerStarted","Data":"46e70fb0bc5976de36fc0679687d84000719b85c5f96fb3f342f63d1a91ff02f"} Oct 11 04:10:57 crc kubenswrapper[4967]: I1011 04:10:57.064911 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4c3cc259-3802-4172-8ae6-446cfbcd4ca5","Type":"ContainerStarted","Data":"2c76affb74be55f25a4ed10d9e62934d07d66c43d12d6b05ab4d84023fcd5735"} Oct 11 04:10:57 crc kubenswrapper[4967]: I1011 04:10:57.064967 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/nova-api-0" event={"ID":"4c3cc259-3802-4172-8ae6-446cfbcd4ca5","Type":"ContainerStarted","Data":"9bdafa75eed72e0816a9e820b6e48b815ff6f5943fea6ff99d8cb05c8f2c12ee"} Oct 11 04:10:57 crc kubenswrapper[4967]: I1011 04:10:57.099127 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-metadata-0" podStartSLOduration=2.09910106 podStartE2EDuration="2.09910106s" podCreationTimestamp="2025-10-11 04:10:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:10:57.08356756 +0000 UTC m=+1185.046776533" watchObservedRunningTime="2025-10-11 04:10:57.09910106 +0000 UTC m=+1185.062310033" Oct 11 04:10:57 crc kubenswrapper[4967]: I1011 04:10:57.123203 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/nova-api-0" podStartSLOduration=2.123172577 podStartE2EDuration="2.123172577s" podCreationTimestamp="2025-10-11 04:10:55 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:10:57.106243025 +0000 UTC m=+1185.069451998" watchObservedRunningTime="2025-10-11 04:10:57.123172577 +0000 UTC m=+1185.086381530" Oct 11 04:10:58 crc kubenswrapper[4967]: I1011 04:10:58.334580 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-scheduler-0" Oct 11 04:11:00 crc kubenswrapper[4967]: I1011 04:11:00.423413 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 11 04:11:00 crc kubenswrapper[4967]: I1011 04:11:00.423734 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-metadata-0" Oct 11 04:11:03 crc kubenswrapper[4967]: I1011 04:11:03.334326 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-scheduler-0" Oct 11 04:11:03 crc kubenswrapper[4967]: I1011 04:11:03.370530 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-scheduler-0" Oct 11 04:11:04 crc kubenswrapper[4967]: I1011 04:11:04.186006 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-scheduler-0" Oct 11 04:11:04 crc kubenswrapper[4967]: I1011 04:11:04.329155 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ceilometer-0" Oct 11 04:11:05 crc kubenswrapper[4967]: I1011 04:11:05.423683 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 11 04:11:05 crc kubenswrapper[4967]: I1011 04:11:05.424224 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-metadata-0" Oct 11 04:11:05 crc kubenswrapper[4967]: I1011 04:11:05.455615 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 11 04:11:05 crc kubenswrapper[4967]: I1011 04:11:05.455686 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openstack/nova-api-0" Oct 11 04:11:06 crc kubenswrapper[4967]: I1011 04:11:06.437253 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8e701cd5-1cf9-4ed0-87c6-2023c41e7982" containerName="nova-metadata-metadata" probeResult="failure" output="Get \"https://10.217.0.212:8775/\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Oct 11 04:11:06 crc kubenswrapper[4967]: I1011 04:11:06.437273 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-metadata-0" podUID="8e701cd5-1cf9-4ed0-87c6-2023c41e7982" containerName="nova-metadata-log" probeResult="failure" output="Get \"https://10.217.0.212:8775/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 11 04:11:06 crc kubenswrapper[4967]: I1011 04:11:06.473230 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4c3cc259-3802-4172-8ae6-446cfbcd4ca5" containerName="nova-api-api" probeResult="failure" output="Get \"https://10.217.0.213:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 11 04:11:06 crc kubenswrapper[4967]: I1011 04:11:06.473304 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openstack/nova-api-0" podUID="4c3cc259-3802-4172-8ae6-446cfbcd4ca5" containerName="nova-api-log" probeResult="failure" output="Get \"https://10.217.0.213:8774/\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)" Oct 11 04:11:15 crc kubenswrapper[4967]: I1011 04:11:15.431616 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 11 04:11:15 crc kubenswrapper[4967]: I1011 04:11:15.433624 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-metadata-0" Oct 11 04:11:15 crc kubenswrapper[4967]: I1011 04:11:15.439754 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 11 04:11:15 crc kubenswrapper[4967]: I1011 04:11:15.469450 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 11 04:11:15 crc kubenswrapper[4967]: I1011 04:11:15.469993 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 11 04:11:15 crc kubenswrapper[4967]: I1011 04:11:15.473621 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openstack/nova-api-0" Oct 11 04:11:15 crc kubenswrapper[4967]: I1011 04:11:15.477203 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 11 04:11:16 crc kubenswrapper[4967]: I1011 04:11:16.271889 4967 generic.go:334] "Generic (PLEG): container finished" podID="64c75ea6-aed6-4bf6-8eef-39e76384e14f" containerID="ad668f94310d78ecf52994e730277185b48ab6fae339ec0352d134f3cf5dc2de" exitCode=0 Oct 11 04:11:16 crc kubenswrapper[4967]: I1011 04:11:16.271965 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"64c75ea6-aed6-4bf6-8eef-39e76384e14f","Type":"ContainerDied","Data":"ad668f94310d78ecf52994e730277185b48ab6fae339ec0352d134f3cf5dc2de"} Oct 11 04:11:16 crc kubenswrapper[4967]: I1011 04:11:16.274451 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/nova-api-0" Oct 11 04:11:16 crc kubenswrapper[4967]: I1011 04:11:16.283037 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-metadata-0" Oct 11 04:11:16 crc kubenswrapper[4967]: I1011 04:11:16.290147 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/nova-api-0" Oct 11 04:11:17 crc kubenswrapper[4967]: I1011 04:11:17.286133 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"64c75ea6-aed6-4bf6-8eef-39e76384e14f","Type":"ContainerStarted","Data":"7c228feb81bc0b52cb4186b4e2f9501dce780eefc9dfd8202a09a98de3a28a2f"} Oct 11 04:11:17 crc kubenswrapper[4967]: I1011 04:11:17.286756 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"64c75ea6-aed6-4bf6-8eef-39e76384e14f","Type":"ContainerStarted","Data":"4914b19265d976fa6956e6dbc21188346708b9ad837cbb6786d751020a0057e1"} Oct 11 04:11:18 crc kubenswrapper[4967]: I1011 04:11:18.303873 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/ironic-conductor-0" event={"ID":"64c75ea6-aed6-4bf6-8eef-39e76384e14f","Type":"ContainerStarted","Data":"b7361923f8290d69686fdfdf6a81e6ed8126befd4209f6fcf0724337b52bc9e7"} Oct 11 04:11:18 crc kubenswrapper[4967]: I1011 04:11:18.304482 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-conductor-0" Oct 11 04:11:18 crc kubenswrapper[4967]: I1011 04:11:18.336757 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/ironic-conductor-0" podStartSLOduration=69.704025221 podStartE2EDuration="2m34.336734795s" podCreationTimestamp="2025-10-11 04:08:44 +0000 UTC" firstStartedPulling="2025-10-11 04:08:47.949977369 +0000 UTC m=+1055.913186302" lastFinishedPulling="2025-10-11 04:10:12.582686933 +0000 UTC m=+1140.545895876" observedRunningTime="2025-10-11 04:11:18.335365734 +0000 UTC m=+1206.298574757" watchObservedRunningTime="2025-10-11 04:11:18.336734795 +0000 UTC m=+1206.299943738" Oct 11 04:11:19 crc kubenswrapper[4967]: I1011 04:11:19.314790 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/ironic-conductor-0" Oct 11 04:11:22 crc kubenswrapper[4967]: I1011 04:11:22.414924 4967 scope.go:117] "RemoveContainer" containerID="275e549b6a7b3950e3c225c5343d97b681645718fe8cbdbcfb1f4325fc5b6b27" Oct 11 04:11:22 crc kubenswrapper[4967]: I1011 04:11:22.450400 4967 scope.go:117] "RemoveContainer" containerID="d02ce1fa87dc4a67713923b2f5f5d220839cbf12d7ef7eb22c9bd771fb3c8920" Oct 11 04:11:22 crc kubenswrapper[4967]: I1011 04:11:22.494825 4967 scope.go:117] "RemoveContainer" containerID="64e21d0eed5b8671f1f847782f3177afa710a544e25e1713597cfa9e2ce9d877" Oct 11 04:11:22 crc kubenswrapper[4967]: I1011 04:11:22.513445 4967 scope.go:117] "RemoveContainer" containerID="f6a5c5ba5a22010cfe674fced11a2d28812f088ff6b617ffd64a731a3a04616a" Oct 11 04:11:22 crc kubenswrapper[4967]: I1011 04:11:22.537540 4967 scope.go:117] "RemoveContainer" containerID="83962bf9b636c5e9588f7254c872df4d327b6b03b40c0521f28e97227e48a89d" Oct 11 04:11:22 crc kubenswrapper[4967]: I1011 04:11:22.609458 4967 scope.go:117] "RemoveContainer" containerID="1c4cb8afea13c57b22e73f4481c68854e398403effb4fc446167bb32a8fd3280" Oct 11 04:11:45 crc kubenswrapper[4967]: I1011 04:11:45.432283 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-conductor-0" Oct 11 04:11:45 crc kubenswrapper[4967]: I1011 04:11:45.433340 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/ironic-conductor-0" Oct 11 04:11:53 crc kubenswrapper[4967]: I1011 04:11:53.623145 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 11 04:11:54 crc kubenswrapper[4967]: I1011 04:11:54.717183 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 11 04:11:57 crc kubenswrapper[4967]: I1011 04:11:57.591043 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-server-0" podUID="3d923c19-a45c-4cb7-9f19-e770c5fe66ed" containerName="rabbitmq" containerID="cri-o://3a0e9dd337a635129967491448fa27c66efed6b3f2cf226b43e39c2a2861a55c" gracePeriod=604797 Oct 11 04:11:58 crc kubenswrapper[4967]: I1011 04:11:58.765871 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openstack/rabbitmq-cell1-server-0" podUID="0c315653-59d1-429f-be9c-c1ac826cff49" containerName="rabbitmq" containerID="cri-o://6954a528ee16848ee7e858a4e86aaad17c994e0a8493821729091c842af849a0" gracePeriod=604796 Oct 11 04:12:00 crc kubenswrapper[4967]: I1011 04:12:00.921962 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-cell1-server-0" podUID="0c315653-59d1-429f-be9c-c1ac826cff49" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.99:5671: connect: connection refused" Oct 11 04:12:01 crc kubenswrapper[4967]: I1011 04:12:01.244985 4967 prober.go:107] "Probe failed" probeType="Readiness" pod="openstack/rabbitmq-server-0" podUID="3d923c19-a45c-4cb7-9f19-e770c5fe66ed" containerName="rabbitmq" probeResult="failure" output="dial tcp 10.217.0.100:5671: connect: connection refused" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.196330 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.300166 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g6lk6\" (UniqueName: \"kubernetes.io/projected/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-kube-api-access-g6lk6\") pod \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.300266 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-pod-info\") pod \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.300333 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-rabbitmq-plugins\") pod \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.300366 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-config-data\") pod \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.300400 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-rabbitmq-confd\") pod \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.300431 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-erlang-cookie-secret\") pod \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.300533 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-rabbitmq-tls\") pod \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.300632 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-rabbitmq-erlang-cookie\") pod \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.300674 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-plugins-conf\") pod \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.300751 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-server-conf\") pod \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.300858 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\" (UID: \"3d923c19-a45c-4cb7-9f19-e770c5fe66ed\") " Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.302203 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "3d923c19-a45c-4cb7-9f19-e770c5fe66ed" (UID: "3d923c19-a45c-4cb7-9f19-e770c5fe66ed"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.303442 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "3d923c19-a45c-4cb7-9f19-e770c5fe66ed" (UID: "3d923c19-a45c-4cb7-9f19-e770c5fe66ed"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.303419 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "3d923c19-a45c-4cb7-9f19-e770c5fe66ed" (UID: "3d923c19-a45c-4cb7-9f19-e770c5fe66ed"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.315480 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "3d923c19-a45c-4cb7-9f19-e770c5fe66ed" (UID: "3d923c19-a45c-4cb7-9f19-e770c5fe66ed"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.315656 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-kube-api-access-g6lk6" (OuterVolumeSpecName: "kube-api-access-g6lk6") pod "3d923c19-a45c-4cb7-9f19-e770c5fe66ed" (UID: "3d923c19-a45c-4cb7-9f19-e770c5fe66ed"). InnerVolumeSpecName "kube-api-access-g6lk6". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.315893 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage10-crc" (OuterVolumeSpecName: "persistence") pod "3d923c19-a45c-4cb7-9f19-e770c5fe66ed" (UID: "3d923c19-a45c-4cb7-9f19-e770c5fe66ed"). InnerVolumeSpecName "local-storage10-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.320497 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "3d923c19-a45c-4cb7-9f19-e770c5fe66ed" (UID: "3d923c19-a45c-4cb7-9f19-e770c5fe66ed"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.327332 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-pod-info" (OuterVolumeSpecName: "pod-info") pod "3d923c19-a45c-4cb7-9f19-e770c5fe66ed" (UID: "3d923c19-a45c-4cb7-9f19-e770c5fe66ed"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.351348 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-config-data" (OuterVolumeSpecName: "config-data") pod "3d923c19-a45c-4cb7-9f19-e770c5fe66ed" (UID: "3d923c19-a45c-4cb7-9f19-e770c5fe66ed"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.403399 4967 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.403631 4967 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.403642 4967 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.403662 4967 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" " Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.403670 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g6lk6\" (UniqueName: \"kubernetes.io/projected/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-kube-api-access-g6lk6\") on node \"crc\" DevicePath \"\"" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.403679 4967 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-pod-info\") on node \"crc\" DevicePath \"\"" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.403687 4967 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.403694 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.403703 4967 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.417295 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-server-conf" (OuterVolumeSpecName: "server-conf") pod "3d923c19-a45c-4cb7-9f19-e770c5fe66ed" (UID: "3d923c19-a45c-4cb7-9f19-e770c5fe66ed"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.431816 4967 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage10-crc" (UniqueName: "kubernetes.io/local-volume/local-storage10-crc") on node "crc" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.473334 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "3d923c19-a45c-4cb7-9f19-e770c5fe66ed" (UID: "3d923c19-a45c-4cb7-9f19-e770c5fe66ed"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.506769 4967 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-server-conf\") on node \"crc\" DevicePath \"\"" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.506806 4967 reconciler_common.go:293] "Volume detached for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") on node \"crc\" DevicePath \"\"" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.506815 4967 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/3d923c19-a45c-4cb7-9f19-e770c5fe66ed-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.807256 4967 generic.go:334] "Generic (PLEG): container finished" podID="3d923c19-a45c-4cb7-9f19-e770c5fe66ed" containerID="3a0e9dd337a635129967491448fa27c66efed6b3f2cf226b43e39c2a2861a55c" exitCode=0 Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.807309 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3d923c19-a45c-4cb7-9f19-e770c5fe66ed","Type":"ContainerDied","Data":"3a0e9dd337a635129967491448fa27c66efed6b3f2cf226b43e39c2a2861a55c"} Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.807326 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.807346 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"3d923c19-a45c-4cb7-9f19-e770c5fe66ed","Type":"ContainerDied","Data":"b5f4e5a5434cd9517206736b24fd70f5ca589f42fce0f05d46488df83b0a74e0"} Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.807367 4967 scope.go:117] "RemoveContainer" containerID="3a0e9dd337a635129967491448fa27c66efed6b3f2cf226b43e39c2a2861a55c" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.833810 4967 scope.go:117] "RemoveContainer" containerID="90a1eab201a7f4eef8d6e63d4754f3e37c2c7b5857764dda95a5217b63654438" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.852947 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.856775 4967 scope.go:117] "RemoveContainer" containerID="3a0e9dd337a635129967491448fa27c66efed6b3f2cf226b43e39c2a2861a55c" Oct 11 04:12:04 crc kubenswrapper[4967]: E1011 04:12:04.861786 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3a0e9dd337a635129967491448fa27c66efed6b3f2cf226b43e39c2a2861a55c\": container with ID starting with 3a0e9dd337a635129967491448fa27c66efed6b3f2cf226b43e39c2a2861a55c not found: ID does not exist" containerID="3a0e9dd337a635129967491448fa27c66efed6b3f2cf226b43e39c2a2861a55c" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.861845 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3a0e9dd337a635129967491448fa27c66efed6b3f2cf226b43e39c2a2861a55c"} err="failed to get container status \"3a0e9dd337a635129967491448fa27c66efed6b3f2cf226b43e39c2a2861a55c\": rpc error: code = NotFound desc = could not find container \"3a0e9dd337a635129967491448fa27c66efed6b3f2cf226b43e39c2a2861a55c\": container with ID starting with 3a0e9dd337a635129967491448fa27c66efed6b3f2cf226b43e39c2a2861a55c not found: ID does not exist" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.861877 4967 scope.go:117] "RemoveContainer" containerID="90a1eab201a7f4eef8d6e63d4754f3e37c2c7b5857764dda95a5217b63654438" Oct 11 04:12:04 crc kubenswrapper[4967]: E1011 04:12:04.865322 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"90a1eab201a7f4eef8d6e63d4754f3e37c2c7b5857764dda95a5217b63654438\": container with ID starting with 90a1eab201a7f4eef8d6e63d4754f3e37c2c7b5857764dda95a5217b63654438 not found: ID does not exist" containerID="90a1eab201a7f4eef8d6e63d4754f3e37c2c7b5857764dda95a5217b63654438" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.865373 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"90a1eab201a7f4eef8d6e63d4754f3e37c2c7b5857764dda95a5217b63654438"} err="failed to get container status \"90a1eab201a7f4eef8d6e63d4754f3e37c2c7b5857764dda95a5217b63654438\": rpc error: code = NotFound desc = could not find container \"90a1eab201a7f4eef8d6e63d4754f3e37c2c7b5857764dda95a5217b63654438\": container with ID starting with 90a1eab201a7f4eef8d6e63d4754f3e37c2c7b5857764dda95a5217b63654438 not found: ID does not exist" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.871396 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.880599 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-server-0"] Oct 11 04:12:04 crc kubenswrapper[4967]: E1011 04:12:04.881182 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d923c19-a45c-4cb7-9f19-e770c5fe66ed" containerName="rabbitmq" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.881206 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d923c19-a45c-4cb7-9f19-e770c5fe66ed" containerName="rabbitmq" Oct 11 04:12:04 crc kubenswrapper[4967]: E1011 04:12:04.881227 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3d923c19-a45c-4cb7-9f19-e770c5fe66ed" containerName="setup-container" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.881233 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="3d923c19-a45c-4cb7-9f19-e770c5fe66ed" containerName="setup-container" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.881490 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="3d923c19-a45c-4cb7-9f19-e770c5fe66ed" containerName="rabbitmq" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.882764 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.886569 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-default-user" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.888533 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-svc" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.888852 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-server-conf" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.889157 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-plugins-conf" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.889361 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-config-data" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.889532 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-server-dockercfg-zdwth" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.889722 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-erlang-cookie" Oct 11 04:12:04 crc kubenswrapper[4967]: I1011 04:12:04.901711 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 11 04:12:05 crc kubenswrapper[4967]: E1011 04:12:05.025414 4967 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3d923c19_a45c_4cb7_9f19_e770c5fe66ed.slice/crio-b5f4e5a5434cd9517206736b24fd70f5ca589f42fce0f05d46488df83b0a74e0\": RecentStats: unable to find data in memory cache]" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.026638 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.026683 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.026720 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.026741 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.026781 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab-server-conf\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.026814 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-45gn4\" (UniqueName: \"kubernetes.io/projected/fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab-kube-api-access-45gn4\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.026843 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab-pod-info\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.026883 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.026913 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab-config-data\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.026944 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.026974 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.128329 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab-config-data\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.128652 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.128693 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.128744 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.128768 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.128809 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.128846 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.128886 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab-server-conf\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.128917 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-45gn4\" (UniqueName: \"kubernetes.io/projected/fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab-kube-api-access-45gn4\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.128941 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab-pod-info\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.128984 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.129535 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab-config-data\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.129635 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab-rabbitmq-erlang-cookie\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.130064 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab-rabbitmq-plugins\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.130427 4967 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") device mount path \"/mnt/openstack/pv10\"" pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.130547 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab-plugins-conf\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.130825 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab-server-conf\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.140931 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab-pod-info\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.142736 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab-rabbitmq-confd\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.146556 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab-erlang-cookie-secret\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.148633 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab-rabbitmq-tls\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.152702 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-45gn4\" (UniqueName: \"kubernetes.io/projected/fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab-kube-api-access-45gn4\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.199672 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage10-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage10-crc\") pod \"rabbitmq-server-0\" (UID: \"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab\") " pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.330161 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.367401 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.433052 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c315653-59d1-429f-be9c-c1ac826cff49-config-data\") pod \"0c315653-59d1-429f-be9c-c1ac826cff49\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.433128 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"persistence\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"0c315653-59d1-429f-be9c-c1ac826cff49\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.433159 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0c315653-59d1-429f-be9c-c1ac826cff49-pod-info\") pod \"0c315653-59d1-429f-be9c-c1ac826cff49\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.433194 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0c315653-59d1-429f-be9c-c1ac826cff49-rabbitmq-tls\") pod \"0c315653-59d1-429f-be9c-c1ac826cff49\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.433272 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0c315653-59d1-429f-be9c-c1ac826cff49-rabbitmq-plugins\") pod \"0c315653-59d1-429f-be9c-c1ac826cff49\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.433297 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h2x4x\" (UniqueName: \"kubernetes.io/projected/0c315653-59d1-429f-be9c-c1ac826cff49-kube-api-access-h2x4x\") pod \"0c315653-59d1-429f-be9c-c1ac826cff49\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.433394 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0c315653-59d1-429f-be9c-c1ac826cff49-rabbitmq-erlang-cookie\") pod \"0c315653-59d1-429f-be9c-c1ac826cff49\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.433426 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0c315653-59d1-429f-be9c-c1ac826cff49-server-conf\") pod \"0c315653-59d1-429f-be9c-c1ac826cff49\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.433465 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0c315653-59d1-429f-be9c-c1ac826cff49-rabbitmq-confd\") pod \"0c315653-59d1-429f-be9c-c1ac826cff49\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.433538 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0c315653-59d1-429f-be9c-c1ac826cff49-plugins-conf\") pod \"0c315653-59d1-429f-be9c-c1ac826cff49\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.433570 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0c315653-59d1-429f-be9c-c1ac826cff49-erlang-cookie-secret\") pod \"0c315653-59d1-429f-be9c-c1ac826cff49\" (UID: \"0c315653-59d1-429f-be9c-c1ac826cff49\") " Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.434050 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c315653-59d1-429f-be9c-c1ac826cff49-rabbitmq-erlang-cookie" (OuterVolumeSpecName: "rabbitmq-erlang-cookie") pod "0c315653-59d1-429f-be9c-c1ac826cff49" (UID: "0c315653-59d1-429f-be9c-c1ac826cff49"). InnerVolumeSpecName "rabbitmq-erlang-cookie". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.434225 4967 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/0c315653-59d1-429f-be9c-c1ac826cff49-rabbitmq-erlang-cookie\") on node \"crc\" DevicePath \"\"" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.437875 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c315653-59d1-429f-be9c-c1ac826cff49-plugins-conf" (OuterVolumeSpecName: "plugins-conf") pod "0c315653-59d1-429f-be9c-c1ac826cff49" (UID: "0c315653-59d1-429f-be9c-c1ac826cff49"). InnerVolumeSpecName "plugins-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.438351 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/0c315653-59d1-429f-be9c-c1ac826cff49-rabbitmq-plugins" (OuterVolumeSpecName: "rabbitmq-plugins") pod "0c315653-59d1-429f-be9c-c1ac826cff49" (UID: "0c315653-59d1-429f-be9c-c1ac826cff49"). InnerVolumeSpecName "rabbitmq-plugins". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.442549 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c315653-59d1-429f-be9c-c1ac826cff49-rabbitmq-tls" (OuterVolumeSpecName: "rabbitmq-tls") pod "0c315653-59d1-429f-be9c-c1ac826cff49" (UID: "0c315653-59d1-429f-be9c-c1ac826cff49"). InnerVolumeSpecName "rabbitmq-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.442901 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c315653-59d1-429f-be9c-c1ac826cff49-kube-api-access-h2x4x" (OuterVolumeSpecName: "kube-api-access-h2x4x") pod "0c315653-59d1-429f-be9c-c1ac826cff49" (UID: "0c315653-59d1-429f-be9c-c1ac826cff49"). InnerVolumeSpecName "kube-api-access-h2x4x". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.443998 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/local-volume/local-storage06-crc" (OuterVolumeSpecName: "persistence") pod "0c315653-59d1-429f-be9c-c1ac826cff49" (UID: "0c315653-59d1-429f-be9c-c1ac826cff49"). InnerVolumeSpecName "local-storage06-crc". PluginName "kubernetes.io/local-volume", VolumeGidValue "" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.452108 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0c315653-59d1-429f-be9c-c1ac826cff49-erlang-cookie-secret" (OuterVolumeSpecName: "erlang-cookie-secret") pod "0c315653-59d1-429f-be9c-c1ac826cff49" (UID: "0c315653-59d1-429f-be9c-c1ac826cff49"). InnerVolumeSpecName "erlang-cookie-secret". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.461817 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/downward-api/0c315653-59d1-429f-be9c-c1ac826cff49-pod-info" (OuterVolumeSpecName: "pod-info") pod "0c315653-59d1-429f-be9c-c1ac826cff49" (UID: "0c315653-59d1-429f-be9c-c1ac826cff49"). InnerVolumeSpecName "pod-info". PluginName "kubernetes.io/downward-api", VolumeGidValue "" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.493644 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c315653-59d1-429f-be9c-c1ac826cff49-config-data" (OuterVolumeSpecName: "config-data") pod "0c315653-59d1-429f-be9c-c1ac826cff49" (UID: "0c315653-59d1-429f-be9c-c1ac826cff49"). InnerVolumeSpecName "config-data". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.514295 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0c315653-59d1-429f-be9c-c1ac826cff49-server-conf" (OuterVolumeSpecName: "server-conf") pod "0c315653-59d1-429f-be9c-c1ac826cff49" (UID: "0c315653-59d1-429f-be9c-c1ac826cff49"). InnerVolumeSpecName "server-conf". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.536062 4967 reconciler_common.go:293] "Volume detached for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/0c315653-59d1-429f-be9c-c1ac826cff49-plugins-conf\") on node \"crc\" DevicePath \"\"" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.536221 4967 reconciler_common.go:293] "Volume detached for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/0c315653-59d1-429f-be9c-c1ac826cff49-erlang-cookie-secret\") on node \"crc\" DevicePath \"\"" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.544874 4967 reconciler_common.go:293] "Volume detached for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/0c315653-59d1-429f-be9c-c1ac826cff49-config-data\") on node \"crc\" DevicePath \"\"" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.544925 4967 reconciler_common.go:286] "operationExecutor.UnmountDevice started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" " Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.544936 4967 reconciler_common.go:293] "Volume detached for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/0c315653-59d1-429f-be9c-c1ac826cff49-pod-info\") on node \"crc\" DevicePath \"\"" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.544946 4967 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/0c315653-59d1-429f-be9c-c1ac826cff49-rabbitmq-tls\") on node \"crc\" DevicePath \"\"" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.544954 4967 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/0c315653-59d1-429f-be9c-c1ac826cff49-rabbitmq-plugins\") on node \"crc\" DevicePath \"\"" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.544966 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h2x4x\" (UniqueName: \"kubernetes.io/projected/0c315653-59d1-429f-be9c-c1ac826cff49-kube-api-access-h2x4x\") on node \"crc\" DevicePath \"\"" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.544976 4967 reconciler_common.go:293] "Volume detached for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/0c315653-59d1-429f-be9c-c1ac826cff49-server-conf\") on node \"crc\" DevicePath \"\"" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.574932 4967 operation_generator.go:917] UnmountDevice succeeded for volume "local-storage06-crc" (UniqueName: "kubernetes.io/local-volume/local-storage06-crc") on node "crc" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.602670 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0c315653-59d1-429f-be9c-c1ac826cff49-rabbitmq-confd" (OuterVolumeSpecName: "rabbitmq-confd") pod "0c315653-59d1-429f-be9c-c1ac826cff49" (UID: "0c315653-59d1-429f-be9c-c1ac826cff49"). InnerVolumeSpecName "rabbitmq-confd". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.647477 4967 reconciler_common.go:293] "Volume detached for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") on node \"crc\" DevicePath \"\"" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.647667 4967 reconciler_common.go:293] "Volume detached for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/0c315653-59d1-429f-be9c-c1ac826cff49-rabbitmq-confd\") on node \"crc\" DevicePath \"\"" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.818219 4967 generic.go:334] "Generic (PLEG): container finished" podID="0c315653-59d1-429f-be9c-c1ac826cff49" containerID="6954a528ee16848ee7e858a4e86aaad17c994e0a8493821729091c842af849a0" exitCode=0 Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.818293 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0c315653-59d1-429f-be9c-c1ac826cff49","Type":"ContainerDied","Data":"6954a528ee16848ee7e858a4e86aaad17c994e0a8493821729091c842af849a0"} Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.818323 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"0c315653-59d1-429f-be9c-c1ac826cff49","Type":"ContainerDied","Data":"cfe9b07ca4a82285c51651ed38d8d66a562ad2394d3de32e4db0214795030210"} Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.818346 4967 scope.go:117] "RemoveContainer" containerID="6954a528ee16848ee7e858a4e86aaad17c994e0a8493821729091c842af849a0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.818483 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.840308 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-server-0"] Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.860866 4967 scope.go:117] "RemoveContainer" containerID="55e8d168f569f266bd42f0f73e79f782f1b67630b934f6721fb999704f507332" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.868311 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.878786 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.901243 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 11 04:12:05 crc kubenswrapper[4967]: E1011 04:12:05.901763 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c315653-59d1-429f-be9c-c1ac826cff49" containerName="rabbitmq" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.901784 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c315653-59d1-429f-be9c-c1ac826cff49" containerName="rabbitmq" Oct 11 04:12:05 crc kubenswrapper[4967]: E1011 04:12:05.901802 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="0c315653-59d1-429f-be9c-c1ac826cff49" containerName="setup-container" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.901815 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="0c315653-59d1-429f-be9c-c1ac826cff49" containerName="setup-container" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.902120 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="0c315653-59d1-429f-be9c-c1ac826cff49" containerName="rabbitmq" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.903438 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.911241 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-default-user" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.911401 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-erlang-cookie" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.911427 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-plugins-conf" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.911530 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-server-conf" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.911663 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"cert-rabbitmq-cell1-svc" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.912260 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openstack"/"rabbitmq-cell1-config-data" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.912629 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openstack"/"rabbitmq-cell1-server-dockercfg-wmdzf" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.918558 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.951990 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.952035 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c41a3ead-949a-4f76-9a58-0e1b4db0abb5-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.952087 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c41a3ead-949a-4f76-9a58-0e1b4db0abb5-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.952113 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c41a3ead-949a-4f76-9a58-0e1b4db0abb5-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.952163 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c41a3ead-949a-4f76-9a58-0e1b4db0abb5-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.952184 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c41a3ead-949a-4f76-9a58-0e1b4db0abb5-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.952204 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c41a3ead-949a-4f76-9a58-0e1b4db0abb5-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.952260 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c41a3ead-949a-4f76-9a58-0e1b4db0abb5-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.952282 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c41a3ead-949a-4f76-9a58-0e1b4db0abb5-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.952308 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jhxm\" (UniqueName: \"kubernetes.io/projected/c41a3ead-949a-4f76-9a58-0e1b4db0abb5-kube-api-access-5jhxm\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:05 crc kubenswrapper[4967]: I1011 04:12:05.952868 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c41a3ead-949a-4f76-9a58-0e1b4db0abb5-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.027147 4967 scope.go:117] "RemoveContainer" containerID="6954a528ee16848ee7e858a4e86aaad17c994e0a8493821729091c842af849a0" Oct 11 04:12:06 crc kubenswrapper[4967]: E1011 04:12:06.029551 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6954a528ee16848ee7e858a4e86aaad17c994e0a8493821729091c842af849a0\": container with ID starting with 6954a528ee16848ee7e858a4e86aaad17c994e0a8493821729091c842af849a0 not found: ID does not exist" containerID="6954a528ee16848ee7e858a4e86aaad17c994e0a8493821729091c842af849a0" Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.029594 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6954a528ee16848ee7e858a4e86aaad17c994e0a8493821729091c842af849a0"} err="failed to get container status \"6954a528ee16848ee7e858a4e86aaad17c994e0a8493821729091c842af849a0\": rpc error: code = NotFound desc = could not find container \"6954a528ee16848ee7e858a4e86aaad17c994e0a8493821729091c842af849a0\": container with ID starting with 6954a528ee16848ee7e858a4e86aaad17c994e0a8493821729091c842af849a0 not found: ID does not exist" Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.029617 4967 scope.go:117] "RemoveContainer" containerID="55e8d168f569f266bd42f0f73e79f782f1b67630b934f6721fb999704f507332" Oct 11 04:12:06 crc kubenswrapper[4967]: E1011 04:12:06.030093 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"55e8d168f569f266bd42f0f73e79f782f1b67630b934f6721fb999704f507332\": container with ID starting with 55e8d168f569f266bd42f0f73e79f782f1b67630b934f6721fb999704f507332 not found: ID does not exist" containerID="55e8d168f569f266bd42f0f73e79f782f1b67630b934f6721fb999704f507332" Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.030145 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"55e8d168f569f266bd42f0f73e79f782f1b67630b934f6721fb999704f507332"} err="failed to get container status \"55e8d168f569f266bd42f0f73e79f782f1b67630b934f6721fb999704f507332\": rpc error: code = NotFound desc = could not find container \"55e8d168f569f266bd42f0f73e79f782f1b67630b934f6721fb999704f507332\": container with ID starting with 55e8d168f569f266bd42f0f73e79f782f1b67630b934f6721fb999704f507332 not found: ID does not exist" Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.054731 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.054801 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c41a3ead-949a-4f76-9a58-0e1b4db0abb5-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.054838 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c41a3ead-949a-4f76-9a58-0e1b4db0abb5-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.054873 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c41a3ead-949a-4f76-9a58-0e1b4db0abb5-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.054938 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c41a3ead-949a-4f76-9a58-0e1b4db0abb5-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.054959 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c41a3ead-949a-4f76-9a58-0e1b4db0abb5-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.054984 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c41a3ead-949a-4f76-9a58-0e1b4db0abb5-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.055661 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c41a3ead-949a-4f76-9a58-0e1b4db0abb5-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.055707 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c41a3ead-949a-4f76-9a58-0e1b4db0abb5-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.055867 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-erlang-cookie\" (UniqueName: \"kubernetes.io/empty-dir/c41a3ead-949a-4f76-9a58-0e1b4db0abb5-rabbitmq-erlang-cookie\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.056222 4967 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") device mount path \"/mnt/openstack/pv06\"" pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.056272 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-plugins\" (UniqueName: \"kubernetes.io/empty-dir/c41a3ead-949a-4f76-9a58-0e1b4db0abb5-rabbitmq-plugins\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.057026 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"server-conf\" (UniqueName: \"kubernetes.io/configmap/c41a3ead-949a-4f76-9a58-0e1b4db0abb5-server-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.057130 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jhxm\" (UniqueName: \"kubernetes.io/projected/c41a3ead-949a-4f76-9a58-0e1b4db0abb5-kube-api-access-5jhxm\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.057203 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-data\" (UniqueName: \"kubernetes.io/configmap/c41a3ead-949a-4f76-9a58-0e1b4db0abb5-config-data\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.057591 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c41a3ead-949a-4f76-9a58-0e1b4db0abb5-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.058545 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-conf\" (UniqueName: \"kubernetes.io/configmap/c41a3ead-949a-4f76-9a58-0e1b4db0abb5-plugins-conf\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.059996 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-confd\" (UniqueName: \"kubernetes.io/projected/c41a3ead-949a-4f76-9a58-0e1b4db0abb5-rabbitmq-confd\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.060318 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-info\" (UniqueName: \"kubernetes.io/downward-api/c41a3ead-949a-4f76-9a58-0e1b4db0abb5-pod-info\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.060979 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"erlang-cookie-secret\" (UniqueName: \"kubernetes.io/secret/c41a3ead-949a-4f76-9a58-0e1b4db0abb5-erlang-cookie-secret\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.070829 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rabbitmq-tls\" (UniqueName: \"kubernetes.io/projected/c41a3ead-949a-4f76-9a58-0e1b4db0abb5-rabbitmq-tls\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.074954 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jhxm\" (UniqueName: \"kubernetes.io/projected/c41a3ead-949a-4f76-9a58-0e1b4db0abb5-kube-api-access-5jhxm\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.104436 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"local-storage06-crc\" (UniqueName: \"kubernetes.io/local-volume/local-storage06-crc\") pod \"rabbitmq-cell1-server-0\" (UID: \"c41a3ead-949a-4f76-9a58-0e1b4db0abb5\") " pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.196898 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.730115 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openstack/rabbitmq-cell1-server-0"] Oct 11 04:12:06 crc kubenswrapper[4967]: W1011 04:12:06.733652 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podc41a3ead_949a_4f76_9a58_0e1b4db0abb5.slice/crio-2672073a068ab2dd532dc7aeb8955dddb8b5e3855fd84c52f41b9705ecfa0ffd WatchSource:0}: Error finding container 2672073a068ab2dd532dc7aeb8955dddb8b5e3855fd84c52f41b9705ecfa0ffd: Status 404 returned error can't find the container with id 2672073a068ab2dd532dc7aeb8955dddb8b5e3855fd84c52f41b9705ecfa0ffd Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.828558 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0c315653-59d1-429f-be9c-c1ac826cff49" path="/var/lib/kubelet/pods/0c315653-59d1-429f-be9c-c1ac826cff49/volumes" Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.829856 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3d923c19-a45c-4cb7-9f19-e770c5fe66ed" path="/var/lib/kubelet/pods/3d923c19-a45c-4cb7-9f19-e770c5fe66ed/volumes" Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.835461 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c41a3ead-949a-4f76-9a58-0e1b4db0abb5","Type":"ContainerStarted","Data":"2672073a068ab2dd532dc7aeb8955dddb8b5e3855fd84c52f41b9705ecfa0ffd"} Oct 11 04:12:06 crc kubenswrapper[4967]: I1011 04:12:06.836556 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab","Type":"ContainerStarted","Data":"65ae58467baf92fb92d55c67a05cc0dcc5293090fd6055c709c59c4d197ebbb1"} Oct 11 04:12:08 crc kubenswrapper[4967]: I1011 04:12:08.858170 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c41a3ead-949a-4f76-9a58-0e1b4db0abb5","Type":"ContainerStarted","Data":"f2ffa64c5dc77c047f371f09b2245fa517ba560783d85a4efbf679311b4e3f3a"} Oct 11 04:12:08 crc kubenswrapper[4967]: I1011 04:12:08.862116 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab","Type":"ContainerStarted","Data":"c5b15b6bf5b85586c5f0aebc36e26a1f9523ca772f7ace361b3d6ad019623ec2"} Oct 11 04:12:14 crc kubenswrapper[4967]: I1011 04:12:14.089499 4967 patch_prober.go:28] interesting pod/machine-config-daemon-sgrv2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:12:14 crc kubenswrapper[4967]: I1011 04:12:14.090531 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:12:17 crc kubenswrapper[4967]: E1011 04:12:17.663581 4967 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.50:60190->38.102.83.50:40947: write tcp 38.102.83.50:60190->38.102.83.50:40947: write: connection reset by peer Oct 11 04:12:41 crc kubenswrapper[4967]: I1011 04:12:41.237204 4967 generic.go:334] "Generic (PLEG): container finished" podID="c41a3ead-949a-4f76-9a58-0e1b4db0abb5" containerID="f2ffa64c5dc77c047f371f09b2245fa517ba560783d85a4efbf679311b4e3f3a" exitCode=0 Oct 11 04:12:41 crc kubenswrapper[4967]: I1011 04:12:41.237490 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c41a3ead-949a-4f76-9a58-0e1b4db0abb5","Type":"ContainerDied","Data":"f2ffa64c5dc77c047f371f09b2245fa517ba560783d85a4efbf679311b4e3f3a"} Oct 11 04:12:41 crc kubenswrapper[4967]: I1011 04:12:41.242089 4967 generic.go:334] "Generic (PLEG): container finished" podID="fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab" containerID="c5b15b6bf5b85586c5f0aebc36e26a1f9523ca772f7ace361b3d6ad019623ec2" exitCode=0 Oct 11 04:12:41 crc kubenswrapper[4967]: I1011 04:12:41.242143 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab","Type":"ContainerDied","Data":"c5b15b6bf5b85586c5f0aebc36e26a1f9523ca772f7ace361b3d6ad019623ec2"} Oct 11 04:12:42 crc kubenswrapper[4967]: I1011 04:12:42.258782 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-cell1-server-0" event={"ID":"c41a3ead-949a-4f76-9a58-0e1b4db0abb5","Type":"ContainerStarted","Data":"f15ced3b77bfcab2b3a35d291830a56605b000c0bb1375a5b301d52b96731701"} Oct 11 04:12:42 crc kubenswrapper[4967]: I1011 04:12:42.261682 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:12:42 crc kubenswrapper[4967]: I1011 04:12:42.267657 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openstack/rabbitmq-server-0" event={"ID":"fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab","Type":"ContainerStarted","Data":"1002c809612d5b6dcc2774be888de2451ff7839174c946089ce2eb11262c2a11"} Oct 11 04:12:42 crc kubenswrapper[4967]: I1011 04:12:42.272874 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openstack/rabbitmq-server-0" Oct 11 04:12:42 crc kubenswrapper[4967]: I1011 04:12:42.305055 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-cell1-server-0" podStartSLOduration=37.305034081 podStartE2EDuration="37.305034081s" podCreationTimestamp="2025-10-11 04:12:05 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:12:42.293599927 +0000 UTC m=+1290.256808920" watchObservedRunningTime="2025-10-11 04:12:42.305034081 +0000 UTC m=+1290.268243014" Oct 11 04:12:44 crc kubenswrapper[4967]: I1011 04:12:44.088770 4967 patch_prober.go:28] interesting pod/machine-config-daemon-sgrv2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:12:44 crc kubenswrapper[4967]: I1011 04:12:44.089235 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:12:55 crc kubenswrapper[4967]: I1011 04:12:55.336370 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-server-0" Oct 11 04:12:55 crc kubenswrapper[4967]: I1011 04:12:55.373123 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openstack/rabbitmq-server-0" podStartSLOduration=51.37310321 podStartE2EDuration="51.37310321s" podCreationTimestamp="2025-10-11 04:12:04 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:12:42.324605725 +0000 UTC m=+1290.287814688" watchObservedRunningTime="2025-10-11 04:12:55.37310321 +0000 UTC m=+1303.336312153" Oct 11 04:12:56 crc kubenswrapper[4967]: I1011 04:12:56.201584 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openstack/rabbitmq-cell1-server-0" Oct 11 04:13:14 crc kubenswrapper[4967]: I1011 04:13:14.089800 4967 patch_prober.go:28] interesting pod/machine-config-daemon-sgrv2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:13:14 crc kubenswrapper[4967]: I1011 04:13:14.090675 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:13:14 crc kubenswrapper[4967]: I1011 04:13:14.090773 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" Oct 11 04:13:14 crc kubenswrapper[4967]: I1011 04:13:14.092386 4967 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"c14dde3bbb545266e69254f79fceffdb0e7aa40ff2dad2c747eae9612aaecda9"} pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 11 04:13:14 crc kubenswrapper[4967]: I1011 04:13:14.092558 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" containerID="cri-o://c14dde3bbb545266e69254f79fceffdb0e7aa40ff2dad2c747eae9612aaecda9" gracePeriod=600 Oct 11 04:13:14 crc kubenswrapper[4967]: I1011 04:13:14.647353 4967 generic.go:334] "Generic (PLEG): container finished" podID="97c07678-14be-410c-b61f-498cb49bc960" containerID="c14dde3bbb545266e69254f79fceffdb0e7aa40ff2dad2c747eae9612aaecda9" exitCode=0 Oct 11 04:13:14 crc kubenswrapper[4967]: I1011 04:13:14.647419 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" event={"ID":"97c07678-14be-410c-b61f-498cb49bc960","Type":"ContainerDied","Data":"c14dde3bbb545266e69254f79fceffdb0e7aa40ff2dad2c747eae9612aaecda9"} Oct 11 04:13:14 crc kubenswrapper[4967]: I1011 04:13:14.647763 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" event={"ID":"97c07678-14be-410c-b61f-498cb49bc960","Type":"ContainerStarted","Data":"6010a3358387442fdc0ff1dfa8bf2d3fb285ef11cfb5f18ad38fe86548ca1f20"} Oct 11 04:13:14 crc kubenswrapper[4967]: I1011 04:13:14.647799 4967 scope.go:117] "RemoveContainer" containerID="37a6f451ae95b67f0b65e3cc023ede1c1770c6a148d14dd88cd2fe303a4c8469" Oct 11 04:13:22 crc kubenswrapper[4967]: I1011 04:13:22.900038 4967 scope.go:117] "RemoveContainer" containerID="f8dc539c1c9ddd669160c80b33dbd71e7b11e50401ace2b49135c047d6adf682" Oct 11 04:14:21 crc kubenswrapper[4967]: I1011 04:14:21.874343 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-srxcl"] Oct 11 04:14:21 crc kubenswrapper[4967]: I1011 04:14:21.876966 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-srxcl" Oct 11 04:14:21 crc kubenswrapper[4967]: I1011 04:14:21.884409 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-srxcl"] Oct 11 04:14:21 crc kubenswrapper[4967]: I1011 04:14:21.964083 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b96b197c-2927-4f7a-94b2-9282ed56eee6-utilities\") pod \"community-operators-srxcl\" (UID: \"b96b197c-2927-4f7a-94b2-9282ed56eee6\") " pod="openshift-marketplace/community-operators-srxcl" Oct 11 04:14:21 crc kubenswrapper[4967]: I1011 04:14:21.964512 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cmbrh\" (UniqueName: \"kubernetes.io/projected/b96b197c-2927-4f7a-94b2-9282ed56eee6-kube-api-access-cmbrh\") pod \"community-operators-srxcl\" (UID: \"b96b197c-2927-4f7a-94b2-9282ed56eee6\") " pod="openshift-marketplace/community-operators-srxcl" Oct 11 04:14:21 crc kubenswrapper[4967]: I1011 04:14:21.964537 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b96b197c-2927-4f7a-94b2-9282ed56eee6-catalog-content\") pod \"community-operators-srxcl\" (UID: \"b96b197c-2927-4f7a-94b2-9282ed56eee6\") " pod="openshift-marketplace/community-operators-srxcl" Oct 11 04:14:22 crc kubenswrapper[4967]: I1011 04:14:22.066123 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cmbrh\" (UniqueName: \"kubernetes.io/projected/b96b197c-2927-4f7a-94b2-9282ed56eee6-kube-api-access-cmbrh\") pod \"community-operators-srxcl\" (UID: \"b96b197c-2927-4f7a-94b2-9282ed56eee6\") " pod="openshift-marketplace/community-operators-srxcl" Oct 11 04:14:22 crc kubenswrapper[4967]: I1011 04:14:22.066171 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b96b197c-2927-4f7a-94b2-9282ed56eee6-catalog-content\") pod \"community-operators-srxcl\" (UID: \"b96b197c-2927-4f7a-94b2-9282ed56eee6\") " pod="openshift-marketplace/community-operators-srxcl" Oct 11 04:14:22 crc kubenswrapper[4967]: I1011 04:14:22.066224 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b96b197c-2927-4f7a-94b2-9282ed56eee6-utilities\") pod \"community-operators-srxcl\" (UID: \"b96b197c-2927-4f7a-94b2-9282ed56eee6\") " pod="openshift-marketplace/community-operators-srxcl" Oct 11 04:14:22 crc kubenswrapper[4967]: I1011 04:14:22.066647 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b96b197c-2927-4f7a-94b2-9282ed56eee6-utilities\") pod \"community-operators-srxcl\" (UID: \"b96b197c-2927-4f7a-94b2-9282ed56eee6\") " pod="openshift-marketplace/community-operators-srxcl" Oct 11 04:14:22 crc kubenswrapper[4967]: I1011 04:14:22.066742 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b96b197c-2927-4f7a-94b2-9282ed56eee6-catalog-content\") pod \"community-operators-srxcl\" (UID: \"b96b197c-2927-4f7a-94b2-9282ed56eee6\") " pod="openshift-marketplace/community-operators-srxcl" Oct 11 04:14:22 crc kubenswrapper[4967]: I1011 04:14:22.087856 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cmbrh\" (UniqueName: \"kubernetes.io/projected/b96b197c-2927-4f7a-94b2-9282ed56eee6-kube-api-access-cmbrh\") pod \"community-operators-srxcl\" (UID: \"b96b197c-2927-4f7a-94b2-9282ed56eee6\") " pod="openshift-marketplace/community-operators-srxcl" Oct 11 04:14:22 crc kubenswrapper[4967]: I1011 04:14:22.207486 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-srxcl" Oct 11 04:14:22 crc kubenswrapper[4967]: I1011 04:14:22.730163 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-srxcl"] Oct 11 04:14:22 crc kubenswrapper[4967]: I1011 04:14:22.984628 4967 scope.go:117] "RemoveContainer" containerID="7fc00872eb20d715869841e872d53e5203e2c809679a6cf3edc55987d07c4441" Oct 11 04:14:23 crc kubenswrapper[4967]: I1011 04:14:23.010845 4967 scope.go:117] "RemoveContainer" containerID="dea40148c9a1a67da3b4bd3d86f7758d94d9b0d3207421731cb4c1e6dbc02697" Oct 11 04:14:23 crc kubenswrapper[4967]: I1011 04:14:23.060514 4967 scope.go:117] "RemoveContainer" containerID="3c1123635771bf7e931bd9889850148e163ce9c271fedcbabe1846b6549fdd39" Oct 11 04:14:23 crc kubenswrapper[4967]: I1011 04:14:23.085200 4967 scope.go:117] "RemoveContainer" containerID="8d452e5bcdee06ff27ad299bd16449ebf91079000282280714301c18395e32a6" Oct 11 04:14:23 crc kubenswrapper[4967]: I1011 04:14:23.124546 4967 scope.go:117] "RemoveContainer" containerID="e06f2ac89aa25dc24dc2dab0cb337b7fe9eafd14405c4942ed856a86f494bb9d" Oct 11 04:14:23 crc kubenswrapper[4967]: I1011 04:14:23.421433 4967 generic.go:334] "Generic (PLEG): container finished" podID="b96b197c-2927-4f7a-94b2-9282ed56eee6" containerID="7f874edfc01f7954cd30a95d1416dc4f9d7b184698368dc00c3ac98661ef196c" exitCode=0 Oct 11 04:14:23 crc kubenswrapper[4967]: I1011 04:14:23.421512 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-srxcl" event={"ID":"b96b197c-2927-4f7a-94b2-9282ed56eee6","Type":"ContainerDied","Data":"7f874edfc01f7954cd30a95d1416dc4f9d7b184698368dc00c3ac98661ef196c"} Oct 11 04:14:23 crc kubenswrapper[4967]: I1011 04:14:23.422259 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-srxcl" event={"ID":"b96b197c-2927-4f7a-94b2-9282ed56eee6","Type":"ContainerStarted","Data":"d43b50a45cc032ee08c33cecfe8beacf921ead39741f2fc95e402a9e31337835"} Oct 11 04:14:23 crc kubenswrapper[4967]: I1011 04:14:23.427559 4967 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 11 04:14:24 crc kubenswrapper[4967]: I1011 04:14:24.436609 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-srxcl" event={"ID":"b96b197c-2927-4f7a-94b2-9282ed56eee6","Type":"ContainerStarted","Data":"d5b2a03a8aaefbb2e8d246ad9ca9cf813a92c8288298f8d38fb35acaaaca6423"} Oct 11 04:14:25 crc kubenswrapper[4967]: I1011 04:14:25.451559 4967 generic.go:334] "Generic (PLEG): container finished" podID="b96b197c-2927-4f7a-94b2-9282ed56eee6" containerID="d5b2a03a8aaefbb2e8d246ad9ca9cf813a92c8288298f8d38fb35acaaaca6423" exitCode=0 Oct 11 04:14:25 crc kubenswrapper[4967]: I1011 04:14:25.451627 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-srxcl" event={"ID":"b96b197c-2927-4f7a-94b2-9282ed56eee6","Type":"ContainerDied","Data":"d5b2a03a8aaefbb2e8d246ad9ca9cf813a92c8288298f8d38fb35acaaaca6423"} Oct 11 04:14:25 crc kubenswrapper[4967]: I1011 04:14:25.451681 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-srxcl" event={"ID":"b96b197c-2927-4f7a-94b2-9282ed56eee6","Type":"ContainerStarted","Data":"b0104978e98463acf7d3d9a54429dc77675d4aba501b5721d3b7602c51cbf21b"} Oct 11 04:14:32 crc kubenswrapper[4967]: I1011 04:14:32.208644 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-srxcl" Oct 11 04:14:32 crc kubenswrapper[4967]: I1011 04:14:32.209159 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-srxcl" Oct 11 04:14:32 crc kubenswrapper[4967]: I1011 04:14:32.274874 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-srxcl" Oct 11 04:14:32 crc kubenswrapper[4967]: I1011 04:14:32.333679 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-srxcl" podStartSLOduration=9.76908861 podStartE2EDuration="11.333649501s" podCreationTimestamp="2025-10-11 04:14:21 +0000 UTC" firstStartedPulling="2025-10-11 04:14:23.427326417 +0000 UTC m=+1391.390535360" lastFinishedPulling="2025-10-11 04:14:24.991887318 +0000 UTC m=+1392.955096251" observedRunningTime="2025-10-11 04:14:25.478469152 +0000 UTC m=+1393.441678125" watchObservedRunningTime="2025-10-11 04:14:32.333649501 +0000 UTC m=+1400.296858454" Oct 11 04:14:32 crc kubenswrapper[4967]: I1011 04:14:32.586558 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-srxcl" Oct 11 04:14:32 crc kubenswrapper[4967]: I1011 04:14:32.641418 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-srxcl"] Oct 11 04:14:34 crc kubenswrapper[4967]: I1011 04:14:34.553514 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-srxcl" podUID="b96b197c-2927-4f7a-94b2-9282ed56eee6" containerName="registry-server" containerID="cri-o://b0104978e98463acf7d3d9a54429dc77675d4aba501b5721d3b7602c51cbf21b" gracePeriod=2 Oct 11 04:14:35 crc kubenswrapper[4967]: I1011 04:14:35.115651 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-srxcl" Oct 11 04:14:35 crc kubenswrapper[4967]: I1011 04:14:35.149309 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cmbrh\" (UniqueName: \"kubernetes.io/projected/b96b197c-2927-4f7a-94b2-9282ed56eee6-kube-api-access-cmbrh\") pod \"b96b197c-2927-4f7a-94b2-9282ed56eee6\" (UID: \"b96b197c-2927-4f7a-94b2-9282ed56eee6\") " Oct 11 04:14:35 crc kubenswrapper[4967]: I1011 04:14:35.149500 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b96b197c-2927-4f7a-94b2-9282ed56eee6-catalog-content\") pod \"b96b197c-2927-4f7a-94b2-9282ed56eee6\" (UID: \"b96b197c-2927-4f7a-94b2-9282ed56eee6\") " Oct 11 04:14:35 crc kubenswrapper[4967]: I1011 04:14:35.149561 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b96b197c-2927-4f7a-94b2-9282ed56eee6-utilities\") pod \"b96b197c-2927-4f7a-94b2-9282ed56eee6\" (UID: \"b96b197c-2927-4f7a-94b2-9282ed56eee6\") " Oct 11 04:14:35 crc kubenswrapper[4967]: I1011 04:14:35.150855 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b96b197c-2927-4f7a-94b2-9282ed56eee6-utilities" (OuterVolumeSpecName: "utilities") pod "b96b197c-2927-4f7a-94b2-9282ed56eee6" (UID: "b96b197c-2927-4f7a-94b2-9282ed56eee6"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:14:35 crc kubenswrapper[4967]: I1011 04:14:35.155220 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b96b197c-2927-4f7a-94b2-9282ed56eee6-kube-api-access-cmbrh" (OuterVolumeSpecName: "kube-api-access-cmbrh") pod "b96b197c-2927-4f7a-94b2-9282ed56eee6" (UID: "b96b197c-2927-4f7a-94b2-9282ed56eee6"). InnerVolumeSpecName "kube-api-access-cmbrh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:14:35 crc kubenswrapper[4967]: I1011 04:14:35.206234 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b96b197c-2927-4f7a-94b2-9282ed56eee6-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b96b197c-2927-4f7a-94b2-9282ed56eee6" (UID: "b96b197c-2927-4f7a-94b2-9282ed56eee6"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:14:35 crc kubenswrapper[4967]: I1011 04:14:35.251974 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cmbrh\" (UniqueName: \"kubernetes.io/projected/b96b197c-2927-4f7a-94b2-9282ed56eee6-kube-api-access-cmbrh\") on node \"crc\" DevicePath \"\"" Oct 11 04:14:35 crc kubenswrapper[4967]: I1011 04:14:35.252005 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b96b197c-2927-4f7a-94b2-9282ed56eee6-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 04:14:35 crc kubenswrapper[4967]: I1011 04:14:35.252014 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b96b197c-2927-4f7a-94b2-9282ed56eee6-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 04:14:35 crc kubenswrapper[4967]: I1011 04:14:35.569558 4967 generic.go:334] "Generic (PLEG): container finished" podID="b96b197c-2927-4f7a-94b2-9282ed56eee6" containerID="b0104978e98463acf7d3d9a54429dc77675d4aba501b5721d3b7602c51cbf21b" exitCode=0 Oct 11 04:14:35 crc kubenswrapper[4967]: I1011 04:14:35.569621 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-srxcl" event={"ID":"b96b197c-2927-4f7a-94b2-9282ed56eee6","Type":"ContainerDied","Data":"b0104978e98463acf7d3d9a54429dc77675d4aba501b5721d3b7602c51cbf21b"} Oct 11 04:14:35 crc kubenswrapper[4967]: I1011 04:14:35.569670 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-srxcl" Oct 11 04:14:35 crc kubenswrapper[4967]: I1011 04:14:35.569699 4967 scope.go:117] "RemoveContainer" containerID="b0104978e98463acf7d3d9a54429dc77675d4aba501b5721d3b7602c51cbf21b" Oct 11 04:14:35 crc kubenswrapper[4967]: I1011 04:14:35.569681 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-srxcl" event={"ID":"b96b197c-2927-4f7a-94b2-9282ed56eee6","Type":"ContainerDied","Data":"d43b50a45cc032ee08c33cecfe8beacf921ead39741f2fc95e402a9e31337835"} Oct 11 04:14:35 crc kubenswrapper[4967]: I1011 04:14:35.596314 4967 scope.go:117] "RemoveContainer" containerID="d5b2a03a8aaefbb2e8d246ad9ca9cf813a92c8288298f8d38fb35acaaaca6423" Oct 11 04:14:35 crc kubenswrapper[4967]: I1011 04:14:35.617307 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-srxcl"] Oct 11 04:14:35 crc kubenswrapper[4967]: I1011 04:14:35.626264 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-srxcl"] Oct 11 04:14:35 crc kubenswrapper[4967]: I1011 04:14:35.635703 4967 scope.go:117] "RemoveContainer" containerID="7f874edfc01f7954cd30a95d1416dc4f9d7b184698368dc00c3ac98661ef196c" Oct 11 04:14:35 crc kubenswrapper[4967]: I1011 04:14:35.677252 4967 scope.go:117] "RemoveContainer" containerID="b0104978e98463acf7d3d9a54429dc77675d4aba501b5721d3b7602c51cbf21b" Oct 11 04:14:35 crc kubenswrapper[4967]: E1011 04:14:35.677757 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b0104978e98463acf7d3d9a54429dc77675d4aba501b5721d3b7602c51cbf21b\": container with ID starting with b0104978e98463acf7d3d9a54429dc77675d4aba501b5721d3b7602c51cbf21b not found: ID does not exist" containerID="b0104978e98463acf7d3d9a54429dc77675d4aba501b5721d3b7602c51cbf21b" Oct 11 04:14:35 crc kubenswrapper[4967]: I1011 04:14:35.677811 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b0104978e98463acf7d3d9a54429dc77675d4aba501b5721d3b7602c51cbf21b"} err="failed to get container status \"b0104978e98463acf7d3d9a54429dc77675d4aba501b5721d3b7602c51cbf21b\": rpc error: code = NotFound desc = could not find container \"b0104978e98463acf7d3d9a54429dc77675d4aba501b5721d3b7602c51cbf21b\": container with ID starting with b0104978e98463acf7d3d9a54429dc77675d4aba501b5721d3b7602c51cbf21b not found: ID does not exist" Oct 11 04:14:35 crc kubenswrapper[4967]: I1011 04:14:35.677844 4967 scope.go:117] "RemoveContainer" containerID="d5b2a03a8aaefbb2e8d246ad9ca9cf813a92c8288298f8d38fb35acaaaca6423" Oct 11 04:14:35 crc kubenswrapper[4967]: E1011 04:14:35.678404 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d5b2a03a8aaefbb2e8d246ad9ca9cf813a92c8288298f8d38fb35acaaaca6423\": container with ID starting with d5b2a03a8aaefbb2e8d246ad9ca9cf813a92c8288298f8d38fb35acaaaca6423 not found: ID does not exist" containerID="d5b2a03a8aaefbb2e8d246ad9ca9cf813a92c8288298f8d38fb35acaaaca6423" Oct 11 04:14:35 crc kubenswrapper[4967]: I1011 04:14:35.678471 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d5b2a03a8aaefbb2e8d246ad9ca9cf813a92c8288298f8d38fb35acaaaca6423"} err="failed to get container status \"d5b2a03a8aaefbb2e8d246ad9ca9cf813a92c8288298f8d38fb35acaaaca6423\": rpc error: code = NotFound desc = could not find container \"d5b2a03a8aaefbb2e8d246ad9ca9cf813a92c8288298f8d38fb35acaaaca6423\": container with ID starting with d5b2a03a8aaefbb2e8d246ad9ca9cf813a92c8288298f8d38fb35acaaaca6423 not found: ID does not exist" Oct 11 04:14:35 crc kubenswrapper[4967]: I1011 04:14:35.678515 4967 scope.go:117] "RemoveContainer" containerID="7f874edfc01f7954cd30a95d1416dc4f9d7b184698368dc00c3ac98661ef196c" Oct 11 04:14:35 crc kubenswrapper[4967]: E1011 04:14:35.681193 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7f874edfc01f7954cd30a95d1416dc4f9d7b184698368dc00c3ac98661ef196c\": container with ID starting with 7f874edfc01f7954cd30a95d1416dc4f9d7b184698368dc00c3ac98661ef196c not found: ID does not exist" containerID="7f874edfc01f7954cd30a95d1416dc4f9d7b184698368dc00c3ac98661ef196c" Oct 11 04:14:35 crc kubenswrapper[4967]: I1011 04:14:35.681215 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7f874edfc01f7954cd30a95d1416dc4f9d7b184698368dc00c3ac98661ef196c"} err="failed to get container status \"7f874edfc01f7954cd30a95d1416dc4f9d7b184698368dc00c3ac98661ef196c\": rpc error: code = NotFound desc = could not find container \"7f874edfc01f7954cd30a95d1416dc4f9d7b184698368dc00c3ac98661ef196c\": container with ID starting with 7f874edfc01f7954cd30a95d1416dc4f9d7b184698368dc00c3ac98661ef196c not found: ID does not exist" Oct 11 04:14:36 crc kubenswrapper[4967]: I1011 04:14:36.828282 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b96b197c-2927-4f7a-94b2-9282ed56eee6" path="/var/lib/kubelet/pods/b96b197c-2927-4f7a-94b2-9282ed56eee6/volumes" Oct 11 04:14:46 crc kubenswrapper[4967]: I1011 04:14:46.860057 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-lbw2f"] Oct 11 04:14:46 crc kubenswrapper[4967]: E1011 04:14:46.861623 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b96b197c-2927-4f7a-94b2-9282ed56eee6" containerName="extract-utilities" Oct 11 04:14:46 crc kubenswrapper[4967]: I1011 04:14:46.861653 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="b96b197c-2927-4f7a-94b2-9282ed56eee6" containerName="extract-utilities" Oct 11 04:14:46 crc kubenswrapper[4967]: E1011 04:14:46.861683 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b96b197c-2927-4f7a-94b2-9282ed56eee6" containerName="extract-content" Oct 11 04:14:46 crc kubenswrapper[4967]: I1011 04:14:46.861696 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="b96b197c-2927-4f7a-94b2-9282ed56eee6" containerName="extract-content" Oct 11 04:14:46 crc kubenswrapper[4967]: E1011 04:14:46.861742 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b96b197c-2927-4f7a-94b2-9282ed56eee6" containerName="registry-server" Oct 11 04:14:46 crc kubenswrapper[4967]: I1011 04:14:46.861757 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="b96b197c-2927-4f7a-94b2-9282ed56eee6" containerName="registry-server" Oct 11 04:14:46 crc kubenswrapper[4967]: I1011 04:14:46.862175 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="b96b197c-2927-4f7a-94b2-9282ed56eee6" containerName="registry-server" Oct 11 04:14:46 crc kubenswrapper[4967]: I1011 04:14:46.865128 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lbw2f" Oct 11 04:14:46 crc kubenswrapper[4967]: I1011 04:14:46.875000 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lbw2f"] Oct 11 04:14:46 crc kubenswrapper[4967]: I1011 04:14:46.949561 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hskkw\" (UniqueName: \"kubernetes.io/projected/35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec-kube-api-access-hskkw\") pod \"redhat-marketplace-lbw2f\" (UID: \"35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec\") " pod="openshift-marketplace/redhat-marketplace-lbw2f" Oct 11 04:14:46 crc kubenswrapper[4967]: I1011 04:14:46.949623 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec-utilities\") pod \"redhat-marketplace-lbw2f\" (UID: \"35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec\") " pod="openshift-marketplace/redhat-marketplace-lbw2f" Oct 11 04:14:46 crc kubenswrapper[4967]: I1011 04:14:46.949875 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec-catalog-content\") pod \"redhat-marketplace-lbw2f\" (UID: \"35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec\") " pod="openshift-marketplace/redhat-marketplace-lbw2f" Oct 11 04:14:47 crc kubenswrapper[4967]: I1011 04:14:47.052225 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hskkw\" (UniqueName: \"kubernetes.io/projected/35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec-kube-api-access-hskkw\") pod \"redhat-marketplace-lbw2f\" (UID: \"35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec\") " pod="openshift-marketplace/redhat-marketplace-lbw2f" Oct 11 04:14:47 crc kubenswrapper[4967]: I1011 04:14:47.052288 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec-utilities\") pod \"redhat-marketplace-lbw2f\" (UID: \"35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec\") " pod="openshift-marketplace/redhat-marketplace-lbw2f" Oct 11 04:14:47 crc kubenswrapper[4967]: I1011 04:14:47.052404 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec-catalog-content\") pod \"redhat-marketplace-lbw2f\" (UID: \"35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec\") " pod="openshift-marketplace/redhat-marketplace-lbw2f" Oct 11 04:14:47 crc kubenswrapper[4967]: I1011 04:14:47.052902 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec-utilities\") pod \"redhat-marketplace-lbw2f\" (UID: \"35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec\") " pod="openshift-marketplace/redhat-marketplace-lbw2f" Oct 11 04:14:47 crc kubenswrapper[4967]: I1011 04:14:47.052977 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec-catalog-content\") pod \"redhat-marketplace-lbw2f\" (UID: \"35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec\") " pod="openshift-marketplace/redhat-marketplace-lbw2f" Oct 11 04:14:47 crc kubenswrapper[4967]: I1011 04:14:47.072890 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hskkw\" (UniqueName: \"kubernetes.io/projected/35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec-kube-api-access-hskkw\") pod \"redhat-marketplace-lbw2f\" (UID: \"35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec\") " pod="openshift-marketplace/redhat-marketplace-lbw2f" Oct 11 04:14:47 crc kubenswrapper[4967]: I1011 04:14:47.192118 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lbw2f" Oct 11 04:14:47 crc kubenswrapper[4967]: I1011 04:14:47.685963 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-lbw2f"] Oct 11 04:14:47 crc kubenswrapper[4967]: I1011 04:14:47.724114 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbw2f" event={"ID":"35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec","Type":"ContainerStarted","Data":"98c0fc89202f84c754d403f9fa1d09ee312431494bc1b36cde0f996a3874c116"} Oct 11 04:14:48 crc kubenswrapper[4967]: I1011 04:14:48.738522 4967 generic.go:334] "Generic (PLEG): container finished" podID="35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec" containerID="53e1ec2bd07d6825ade020d328ab0ecf6a24d32e7d433741240b3bef94134df2" exitCode=0 Oct 11 04:14:48 crc kubenswrapper[4967]: I1011 04:14:48.738822 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbw2f" event={"ID":"35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec","Type":"ContainerDied","Data":"53e1ec2bd07d6825ade020d328ab0ecf6a24d32e7d433741240b3bef94134df2"} Oct 11 04:14:49 crc kubenswrapper[4967]: I1011 04:14:49.749723 4967 generic.go:334] "Generic (PLEG): container finished" podID="35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec" containerID="ed971888d8280221a15c6cf678c746f52c7f2b5e4723f47bca44535f1580c999" exitCode=0 Oct 11 04:14:49 crc kubenswrapper[4967]: I1011 04:14:49.749849 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbw2f" event={"ID":"35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec","Type":"ContainerDied","Data":"ed971888d8280221a15c6cf678c746f52c7f2b5e4723f47bca44535f1580c999"} Oct 11 04:14:50 crc kubenswrapper[4967]: I1011 04:14:50.764578 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbw2f" event={"ID":"35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec","Type":"ContainerStarted","Data":"fc4b7fbdf42bbd8f48419bb8a790e264777c8ed5d1fdb5a964f06edd75174b1e"} Oct 11 04:14:50 crc kubenswrapper[4967]: I1011 04:14:50.784574 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-lbw2f" podStartSLOduration=3.3334258979999998 podStartE2EDuration="4.784556419s" podCreationTimestamp="2025-10-11 04:14:46 +0000 UTC" firstStartedPulling="2025-10-11 04:14:48.742765999 +0000 UTC m=+1416.705974942" lastFinishedPulling="2025-10-11 04:14:50.19389649 +0000 UTC m=+1418.157105463" observedRunningTime="2025-10-11 04:14:50.782726538 +0000 UTC m=+1418.745935471" watchObservedRunningTime="2025-10-11 04:14:50.784556419 +0000 UTC m=+1418.747765362" Oct 11 04:14:57 crc kubenswrapper[4967]: I1011 04:14:57.192898 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-lbw2f" Oct 11 04:14:57 crc kubenswrapper[4967]: I1011 04:14:57.193460 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-lbw2f" Oct 11 04:14:57 crc kubenswrapper[4967]: I1011 04:14:57.268923 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-lbw2f" Oct 11 04:14:57 crc kubenswrapper[4967]: I1011 04:14:57.907550 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-lbw2f" Oct 11 04:14:57 crc kubenswrapper[4967]: I1011 04:14:57.979195 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lbw2f"] Oct 11 04:14:59 crc kubenswrapper[4967]: I1011 04:14:59.861068 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-lbw2f" podUID="35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec" containerName="registry-server" containerID="cri-o://fc4b7fbdf42bbd8f48419bb8a790e264777c8ed5d1fdb5a964f06edd75174b1e" gracePeriod=2 Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.144856 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335935-src4n"] Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.146750 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-src4n" Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.148799 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.149520 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.156761 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335935-src4n"] Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.239387 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qblpw\" (UniqueName: \"kubernetes.io/projected/a16f74ee-c779-4fd9-b945-3e455927298a-kube-api-access-qblpw\") pod \"collect-profiles-29335935-src4n\" (UID: \"a16f74ee-c779-4fd9-b945-3e455927298a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-src4n" Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.239944 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a16f74ee-c779-4fd9-b945-3e455927298a-config-volume\") pod \"collect-profiles-29335935-src4n\" (UID: \"a16f74ee-c779-4fd9-b945-3e455927298a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-src4n" Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.240022 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a16f74ee-c779-4fd9-b945-3e455927298a-secret-volume\") pod \"collect-profiles-29335935-src4n\" (UID: \"a16f74ee-c779-4fd9-b945-3e455927298a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-src4n" Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.341950 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qblpw\" (UniqueName: \"kubernetes.io/projected/a16f74ee-c779-4fd9-b945-3e455927298a-kube-api-access-qblpw\") pod \"collect-profiles-29335935-src4n\" (UID: \"a16f74ee-c779-4fd9-b945-3e455927298a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-src4n" Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.342040 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a16f74ee-c779-4fd9-b945-3e455927298a-config-volume\") pod \"collect-profiles-29335935-src4n\" (UID: \"a16f74ee-c779-4fd9-b945-3e455927298a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-src4n" Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.342095 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a16f74ee-c779-4fd9-b945-3e455927298a-secret-volume\") pod \"collect-profiles-29335935-src4n\" (UID: \"a16f74ee-c779-4fd9-b945-3e455927298a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-src4n" Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.342929 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a16f74ee-c779-4fd9-b945-3e455927298a-config-volume\") pod \"collect-profiles-29335935-src4n\" (UID: \"a16f74ee-c779-4fd9-b945-3e455927298a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-src4n" Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.349699 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a16f74ee-c779-4fd9-b945-3e455927298a-secret-volume\") pod \"collect-profiles-29335935-src4n\" (UID: \"a16f74ee-c779-4fd9-b945-3e455927298a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-src4n" Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.359210 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qblpw\" (UniqueName: \"kubernetes.io/projected/a16f74ee-c779-4fd9-b945-3e455927298a-kube-api-access-qblpw\") pod \"collect-profiles-29335935-src4n\" (UID: \"a16f74ee-c779-4fd9-b945-3e455927298a\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-src4n" Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.440418 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lbw2f" Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.464474 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-src4n" Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.544291 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec-utilities\") pod \"35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec\" (UID: \"35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec\") " Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.544401 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hskkw\" (UniqueName: \"kubernetes.io/projected/35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec-kube-api-access-hskkw\") pod \"35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec\" (UID: \"35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec\") " Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.544570 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec-catalog-content\") pod \"35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec\" (UID: \"35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec\") " Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.546432 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec-utilities" (OuterVolumeSpecName: "utilities") pod "35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec" (UID: "35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.548246 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec-kube-api-access-hskkw" (OuterVolumeSpecName: "kube-api-access-hskkw") pod "35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec" (UID: "35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec"). InnerVolumeSpecName "kube-api-access-hskkw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.557260 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec" (UID: "35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.648654 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hskkw\" (UniqueName: \"kubernetes.io/projected/35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec-kube-api-access-hskkw\") on node \"crc\" DevicePath \"\"" Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.648989 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.649000 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.872724 4967 generic.go:334] "Generic (PLEG): container finished" podID="35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec" containerID="fc4b7fbdf42bbd8f48419bb8a790e264777c8ed5d1fdb5a964f06edd75174b1e" exitCode=0 Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.872797 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-lbw2f" Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.874317 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbw2f" event={"ID":"35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec","Type":"ContainerDied","Data":"fc4b7fbdf42bbd8f48419bb8a790e264777c8ed5d1fdb5a964f06edd75174b1e"} Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.874525 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-lbw2f" event={"ID":"35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec","Type":"ContainerDied","Data":"98c0fc89202f84c754d403f9fa1d09ee312431494bc1b36cde0f996a3874c116"} Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.874611 4967 scope.go:117] "RemoveContainer" containerID="fc4b7fbdf42bbd8f48419bb8a790e264777c8ed5d1fdb5a964f06edd75174b1e" Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.899579 4967 scope.go:117] "RemoveContainer" containerID="ed971888d8280221a15c6cf678c746f52c7f2b5e4723f47bca44535f1580c999" Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.907312 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-lbw2f"] Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.918717 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-lbw2f"] Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.922395 4967 scope.go:117] "RemoveContainer" containerID="53e1ec2bd07d6825ade020d328ab0ecf6a24d32e7d433741240b3bef94134df2" Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.944088 4967 scope.go:117] "RemoveContainer" containerID="fc4b7fbdf42bbd8f48419bb8a790e264777c8ed5d1fdb5a964f06edd75174b1e" Oct 11 04:15:00 crc kubenswrapper[4967]: E1011 04:15:00.948511 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fc4b7fbdf42bbd8f48419bb8a790e264777c8ed5d1fdb5a964f06edd75174b1e\": container with ID starting with fc4b7fbdf42bbd8f48419bb8a790e264777c8ed5d1fdb5a964f06edd75174b1e not found: ID does not exist" containerID="fc4b7fbdf42bbd8f48419bb8a790e264777c8ed5d1fdb5a964f06edd75174b1e" Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.948756 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fc4b7fbdf42bbd8f48419bb8a790e264777c8ed5d1fdb5a964f06edd75174b1e"} err="failed to get container status \"fc4b7fbdf42bbd8f48419bb8a790e264777c8ed5d1fdb5a964f06edd75174b1e\": rpc error: code = NotFound desc = could not find container \"fc4b7fbdf42bbd8f48419bb8a790e264777c8ed5d1fdb5a964f06edd75174b1e\": container with ID starting with fc4b7fbdf42bbd8f48419bb8a790e264777c8ed5d1fdb5a964f06edd75174b1e not found: ID does not exist" Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.948940 4967 scope.go:117] "RemoveContainer" containerID="ed971888d8280221a15c6cf678c746f52c7f2b5e4723f47bca44535f1580c999" Oct 11 04:15:00 crc kubenswrapper[4967]: E1011 04:15:00.950371 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ed971888d8280221a15c6cf678c746f52c7f2b5e4723f47bca44535f1580c999\": container with ID starting with ed971888d8280221a15c6cf678c746f52c7f2b5e4723f47bca44535f1580c999 not found: ID does not exist" containerID="ed971888d8280221a15c6cf678c746f52c7f2b5e4723f47bca44535f1580c999" Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.950420 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ed971888d8280221a15c6cf678c746f52c7f2b5e4723f47bca44535f1580c999"} err="failed to get container status \"ed971888d8280221a15c6cf678c746f52c7f2b5e4723f47bca44535f1580c999\": rpc error: code = NotFound desc = could not find container \"ed971888d8280221a15c6cf678c746f52c7f2b5e4723f47bca44535f1580c999\": container with ID starting with ed971888d8280221a15c6cf678c746f52c7f2b5e4723f47bca44535f1580c999 not found: ID does not exist" Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.950453 4967 scope.go:117] "RemoveContainer" containerID="53e1ec2bd07d6825ade020d328ab0ecf6a24d32e7d433741240b3bef94134df2" Oct 11 04:15:00 crc kubenswrapper[4967]: E1011 04:15:00.952375 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"53e1ec2bd07d6825ade020d328ab0ecf6a24d32e7d433741240b3bef94134df2\": container with ID starting with 53e1ec2bd07d6825ade020d328ab0ecf6a24d32e7d433741240b3bef94134df2 not found: ID does not exist" containerID="53e1ec2bd07d6825ade020d328ab0ecf6a24d32e7d433741240b3bef94134df2" Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.952469 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"53e1ec2bd07d6825ade020d328ab0ecf6a24d32e7d433741240b3bef94134df2"} err="failed to get container status \"53e1ec2bd07d6825ade020d328ab0ecf6a24d32e7d433741240b3bef94134df2\": rpc error: code = NotFound desc = could not find container \"53e1ec2bd07d6825ade020d328ab0ecf6a24d32e7d433741240b3bef94134df2\": container with ID starting with 53e1ec2bd07d6825ade020d328ab0ecf6a24d32e7d433741240b3bef94134df2 not found: ID does not exist" Oct 11 04:15:00 crc kubenswrapper[4967]: I1011 04:15:00.960370 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335935-src4n"] Oct 11 04:15:00 crc kubenswrapper[4967]: W1011 04:15:00.966227 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda16f74ee_c779_4fd9_b945_3e455927298a.slice/crio-2dee44f77ba0773073f442ce113eac8b6297d906990eae5c3ab88c3f0692314b WatchSource:0}: Error finding container 2dee44f77ba0773073f442ce113eac8b6297d906990eae5c3ab88c3f0692314b: Status 404 returned error can't find the container with id 2dee44f77ba0773073f442ce113eac8b6297d906990eae5c3ab88c3f0692314b Oct 11 04:15:01 crc kubenswrapper[4967]: I1011 04:15:01.891212 4967 generic.go:334] "Generic (PLEG): container finished" podID="a16f74ee-c779-4fd9-b945-3e455927298a" containerID="1d70c62ac174cc377188eeb8da989ccea048761cc91ce73eebc31305076cb34c" exitCode=0 Oct 11 04:15:01 crc kubenswrapper[4967]: I1011 04:15:01.891837 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-src4n" event={"ID":"a16f74ee-c779-4fd9-b945-3e455927298a","Type":"ContainerDied","Data":"1d70c62ac174cc377188eeb8da989ccea048761cc91ce73eebc31305076cb34c"} Oct 11 04:15:01 crc kubenswrapper[4967]: I1011 04:15:01.891909 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-src4n" event={"ID":"a16f74ee-c779-4fd9-b945-3e455927298a","Type":"ContainerStarted","Data":"2dee44f77ba0773073f442ce113eac8b6297d906990eae5c3ab88c3f0692314b"} Oct 11 04:15:02 crc kubenswrapper[4967]: I1011 04:15:02.836784 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec" path="/var/lib/kubelet/pods/35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec/volumes" Oct 11 04:15:02 crc kubenswrapper[4967]: I1011 04:15:02.930781 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-mplb4"] Oct 11 04:15:02 crc kubenswrapper[4967]: E1011 04:15:02.931428 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec" containerName="extract-content" Oct 11 04:15:02 crc kubenswrapper[4967]: I1011 04:15:02.931451 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec" containerName="extract-content" Oct 11 04:15:02 crc kubenswrapper[4967]: E1011 04:15:02.931471 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec" containerName="extract-utilities" Oct 11 04:15:02 crc kubenswrapper[4967]: I1011 04:15:02.931483 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec" containerName="extract-utilities" Oct 11 04:15:02 crc kubenswrapper[4967]: E1011 04:15:02.931550 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec" containerName="registry-server" Oct 11 04:15:02 crc kubenswrapper[4967]: I1011 04:15:02.931563 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec" containerName="registry-server" Oct 11 04:15:02 crc kubenswrapper[4967]: I1011 04:15:02.931882 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="35dd7ae3-4a5f-462d-b73c-09ad90a5c9ec" containerName="registry-server" Oct 11 04:15:02 crc kubenswrapper[4967]: I1011 04:15:02.935344 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mplb4" Oct 11 04:15:02 crc kubenswrapper[4967]: I1011 04:15:02.971541 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mplb4"] Oct 11 04:15:03 crc kubenswrapper[4967]: I1011 04:15:02.999389 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5697bae-d981-4bde-befa-a54c4cd0464b-catalog-content\") pod \"redhat-operators-mplb4\" (UID: \"e5697bae-d981-4bde-befa-a54c4cd0464b\") " pod="openshift-marketplace/redhat-operators-mplb4" Oct 11 04:15:03 crc kubenswrapper[4967]: I1011 04:15:02.999469 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v8zfh\" (UniqueName: \"kubernetes.io/projected/e5697bae-d981-4bde-befa-a54c4cd0464b-kube-api-access-v8zfh\") pod \"redhat-operators-mplb4\" (UID: \"e5697bae-d981-4bde-befa-a54c4cd0464b\") " pod="openshift-marketplace/redhat-operators-mplb4" Oct 11 04:15:03 crc kubenswrapper[4967]: I1011 04:15:02.999530 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5697bae-d981-4bde-befa-a54c4cd0464b-utilities\") pod \"redhat-operators-mplb4\" (UID: \"e5697bae-d981-4bde-befa-a54c4cd0464b\") " pod="openshift-marketplace/redhat-operators-mplb4" Oct 11 04:15:03 crc kubenswrapper[4967]: I1011 04:15:03.101997 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5697bae-d981-4bde-befa-a54c4cd0464b-utilities\") pod \"redhat-operators-mplb4\" (UID: \"e5697bae-d981-4bde-befa-a54c4cd0464b\") " pod="openshift-marketplace/redhat-operators-mplb4" Oct 11 04:15:03 crc kubenswrapper[4967]: I1011 04:15:03.102367 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5697bae-d981-4bde-befa-a54c4cd0464b-catalog-content\") pod \"redhat-operators-mplb4\" (UID: \"e5697bae-d981-4bde-befa-a54c4cd0464b\") " pod="openshift-marketplace/redhat-operators-mplb4" Oct 11 04:15:03 crc kubenswrapper[4967]: I1011 04:15:03.102446 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v8zfh\" (UniqueName: \"kubernetes.io/projected/e5697bae-d981-4bde-befa-a54c4cd0464b-kube-api-access-v8zfh\") pod \"redhat-operators-mplb4\" (UID: \"e5697bae-d981-4bde-befa-a54c4cd0464b\") " pod="openshift-marketplace/redhat-operators-mplb4" Oct 11 04:15:03 crc kubenswrapper[4967]: I1011 04:15:03.103700 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5697bae-d981-4bde-befa-a54c4cd0464b-utilities\") pod \"redhat-operators-mplb4\" (UID: \"e5697bae-d981-4bde-befa-a54c4cd0464b\") " pod="openshift-marketplace/redhat-operators-mplb4" Oct 11 04:15:03 crc kubenswrapper[4967]: I1011 04:15:03.104227 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5697bae-d981-4bde-befa-a54c4cd0464b-catalog-content\") pod \"redhat-operators-mplb4\" (UID: \"e5697bae-d981-4bde-befa-a54c4cd0464b\") " pod="openshift-marketplace/redhat-operators-mplb4" Oct 11 04:15:03 crc kubenswrapper[4967]: I1011 04:15:03.137156 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v8zfh\" (UniqueName: \"kubernetes.io/projected/e5697bae-d981-4bde-befa-a54c4cd0464b-kube-api-access-v8zfh\") pod \"redhat-operators-mplb4\" (UID: \"e5697bae-d981-4bde-befa-a54c4cd0464b\") " pod="openshift-marketplace/redhat-operators-mplb4" Oct 11 04:15:03 crc kubenswrapper[4967]: I1011 04:15:03.278290 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mplb4" Oct 11 04:15:03 crc kubenswrapper[4967]: I1011 04:15:03.371370 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-src4n" Oct 11 04:15:03 crc kubenswrapper[4967]: I1011 04:15:03.513024 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a16f74ee-c779-4fd9-b945-3e455927298a-secret-volume\") pod \"a16f74ee-c779-4fd9-b945-3e455927298a\" (UID: \"a16f74ee-c779-4fd9-b945-3e455927298a\") " Oct 11 04:15:03 crc kubenswrapper[4967]: I1011 04:15:03.513098 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a16f74ee-c779-4fd9-b945-3e455927298a-config-volume\") pod \"a16f74ee-c779-4fd9-b945-3e455927298a\" (UID: \"a16f74ee-c779-4fd9-b945-3e455927298a\") " Oct 11 04:15:03 crc kubenswrapper[4967]: I1011 04:15:03.513342 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qblpw\" (UniqueName: \"kubernetes.io/projected/a16f74ee-c779-4fd9-b945-3e455927298a-kube-api-access-qblpw\") pod \"a16f74ee-c779-4fd9-b945-3e455927298a\" (UID: \"a16f74ee-c779-4fd9-b945-3e455927298a\") " Oct 11 04:15:03 crc kubenswrapper[4967]: I1011 04:15:03.514351 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a16f74ee-c779-4fd9-b945-3e455927298a-config-volume" (OuterVolumeSpecName: "config-volume") pod "a16f74ee-c779-4fd9-b945-3e455927298a" (UID: "a16f74ee-c779-4fd9-b945-3e455927298a"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:15:03 crc kubenswrapper[4967]: I1011 04:15:03.518564 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a16f74ee-c779-4fd9-b945-3e455927298a-kube-api-access-qblpw" (OuterVolumeSpecName: "kube-api-access-qblpw") pod "a16f74ee-c779-4fd9-b945-3e455927298a" (UID: "a16f74ee-c779-4fd9-b945-3e455927298a"). InnerVolumeSpecName "kube-api-access-qblpw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:15:03 crc kubenswrapper[4967]: I1011 04:15:03.519390 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a16f74ee-c779-4fd9-b945-3e455927298a-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "a16f74ee-c779-4fd9-b945-3e455927298a" (UID: "a16f74ee-c779-4fd9-b945-3e455927298a"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:15:03 crc kubenswrapper[4967]: I1011 04:15:03.615602 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qblpw\" (UniqueName: \"kubernetes.io/projected/a16f74ee-c779-4fd9-b945-3e455927298a-kube-api-access-qblpw\") on node \"crc\" DevicePath \"\"" Oct 11 04:15:03 crc kubenswrapper[4967]: I1011 04:15:03.615631 4967 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/a16f74ee-c779-4fd9-b945-3e455927298a-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 11 04:15:03 crc kubenswrapper[4967]: I1011 04:15:03.615642 4967 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/a16f74ee-c779-4fd9-b945-3e455927298a-config-volume\") on node \"crc\" DevicePath \"\"" Oct 11 04:15:03 crc kubenswrapper[4967]: I1011 04:15:03.757802 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-mplb4"] Oct 11 04:15:03 crc kubenswrapper[4967]: I1011 04:15:03.929851 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mplb4" event={"ID":"e5697bae-d981-4bde-befa-a54c4cd0464b","Type":"ContainerStarted","Data":"48ce4db879abc3f783ae7ced47ce5b09f31e0e899d6fffa286e04f94273e9fd5"} Oct 11 04:15:03 crc kubenswrapper[4967]: I1011 04:15:03.929906 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mplb4" event={"ID":"e5697bae-d981-4bde-befa-a54c4cd0464b","Type":"ContainerStarted","Data":"7402c4bd4fb52d770ef5d527ea90e8dec0df585a4800af6c6bf0cbdca21232c6"} Oct 11 04:15:03 crc kubenswrapper[4967]: I1011 04:15:03.935913 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-src4n" event={"ID":"a16f74ee-c779-4fd9-b945-3e455927298a","Type":"ContainerDied","Data":"2dee44f77ba0773073f442ce113eac8b6297d906990eae5c3ab88c3f0692314b"} Oct 11 04:15:03 crc kubenswrapper[4967]: I1011 04:15:03.935951 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2dee44f77ba0773073f442ce113eac8b6297d906990eae5c3ab88c3f0692314b" Oct 11 04:15:03 crc kubenswrapper[4967]: I1011 04:15:03.936004 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335935-src4n" Oct 11 04:15:04 crc kubenswrapper[4967]: I1011 04:15:04.951048 4967 generic.go:334] "Generic (PLEG): container finished" podID="e5697bae-d981-4bde-befa-a54c4cd0464b" containerID="48ce4db879abc3f783ae7ced47ce5b09f31e0e899d6fffa286e04f94273e9fd5" exitCode=0 Oct 11 04:15:04 crc kubenswrapper[4967]: I1011 04:15:04.951165 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mplb4" event={"ID":"e5697bae-d981-4bde-befa-a54c4cd0464b","Type":"ContainerDied","Data":"48ce4db879abc3f783ae7ced47ce5b09f31e0e899d6fffa286e04f94273e9fd5"} Oct 11 04:15:05 crc kubenswrapper[4967]: I1011 04:15:05.965533 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mplb4" event={"ID":"e5697bae-d981-4bde-befa-a54c4cd0464b","Type":"ContainerStarted","Data":"76880dd9c7e69eb10daa9639bcab7d47c49e95ed0706be5991a5ccbcd87cf339"} Oct 11 04:15:08 crc kubenswrapper[4967]: I1011 04:15:08.018532 4967 generic.go:334] "Generic (PLEG): container finished" podID="e5697bae-d981-4bde-befa-a54c4cd0464b" containerID="76880dd9c7e69eb10daa9639bcab7d47c49e95ed0706be5991a5ccbcd87cf339" exitCode=0 Oct 11 04:15:08 crc kubenswrapper[4967]: I1011 04:15:08.018871 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mplb4" event={"ID":"e5697bae-d981-4bde-befa-a54c4cd0464b","Type":"ContainerDied","Data":"76880dd9c7e69eb10daa9639bcab7d47c49e95ed0706be5991a5ccbcd87cf339"} Oct 11 04:15:09 crc kubenswrapper[4967]: I1011 04:15:09.032635 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mplb4" event={"ID":"e5697bae-d981-4bde-befa-a54c4cd0464b","Type":"ContainerStarted","Data":"3f6251b9c53845caf036eb3aa102ef71a72e18250507b0122d9b5a8207d2ddd3"} Oct 11 04:15:13 crc kubenswrapper[4967]: I1011 04:15:13.278495 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-mplb4" Oct 11 04:15:13 crc kubenswrapper[4967]: I1011 04:15:13.279062 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-mplb4" Oct 11 04:15:14 crc kubenswrapper[4967]: I1011 04:15:14.089600 4967 patch_prober.go:28] interesting pod/machine-config-daemon-sgrv2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:15:14 crc kubenswrapper[4967]: I1011 04:15:14.089869 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:15:14 crc kubenswrapper[4967]: I1011 04:15:14.350613 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-mplb4" podUID="e5697bae-d981-4bde-befa-a54c4cd0464b" containerName="registry-server" probeResult="failure" output=< Oct 11 04:15:14 crc kubenswrapper[4967]: timeout: failed to connect service ":50051" within 1s Oct 11 04:15:14 crc kubenswrapper[4967]: > Oct 11 04:15:23 crc kubenswrapper[4967]: I1011 04:15:23.341755 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-mplb4" Oct 11 04:15:23 crc kubenswrapper[4967]: I1011 04:15:23.361711 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-mplb4" podStartSLOduration=17.840475204 podStartE2EDuration="21.361684018s" podCreationTimestamp="2025-10-11 04:15:02 +0000 UTC" firstStartedPulling="2025-10-11 04:15:04.954003291 +0000 UTC m=+1432.917212264" lastFinishedPulling="2025-10-11 04:15:08.475212155 +0000 UTC m=+1436.438421078" observedRunningTime="2025-10-11 04:15:09.056416361 +0000 UTC m=+1437.019625294" watchObservedRunningTime="2025-10-11 04:15:23.361684018 +0000 UTC m=+1451.324892951" Oct 11 04:15:23 crc kubenswrapper[4967]: I1011 04:15:23.408679 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-mplb4" Oct 11 04:15:23 crc kubenswrapper[4967]: I1011 04:15:23.575592 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mplb4"] Oct 11 04:15:25 crc kubenswrapper[4967]: I1011 04:15:25.185923 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-mplb4" podUID="e5697bae-d981-4bde-befa-a54c4cd0464b" containerName="registry-server" containerID="cri-o://3f6251b9c53845caf036eb3aa102ef71a72e18250507b0122d9b5a8207d2ddd3" gracePeriod=2 Oct 11 04:15:25 crc kubenswrapper[4967]: I1011 04:15:25.711219 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mplb4" Oct 11 04:15:25 crc kubenswrapper[4967]: I1011 04:15:25.763688 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5697bae-d981-4bde-befa-a54c4cd0464b-catalog-content\") pod \"e5697bae-d981-4bde-befa-a54c4cd0464b\" (UID: \"e5697bae-d981-4bde-befa-a54c4cd0464b\") " Oct 11 04:15:25 crc kubenswrapper[4967]: I1011 04:15:25.763749 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5697bae-d981-4bde-befa-a54c4cd0464b-utilities\") pod \"e5697bae-d981-4bde-befa-a54c4cd0464b\" (UID: \"e5697bae-d981-4bde-befa-a54c4cd0464b\") " Oct 11 04:15:25 crc kubenswrapper[4967]: I1011 04:15:25.763807 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v8zfh\" (UniqueName: \"kubernetes.io/projected/e5697bae-d981-4bde-befa-a54c4cd0464b-kube-api-access-v8zfh\") pod \"e5697bae-d981-4bde-befa-a54c4cd0464b\" (UID: \"e5697bae-d981-4bde-befa-a54c4cd0464b\") " Oct 11 04:15:25 crc kubenswrapper[4967]: I1011 04:15:25.765018 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5697bae-d981-4bde-befa-a54c4cd0464b-utilities" (OuterVolumeSpecName: "utilities") pod "e5697bae-d981-4bde-befa-a54c4cd0464b" (UID: "e5697bae-d981-4bde-befa-a54c4cd0464b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:15:25 crc kubenswrapper[4967]: I1011 04:15:25.770833 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e5697bae-d981-4bde-befa-a54c4cd0464b-kube-api-access-v8zfh" (OuterVolumeSpecName: "kube-api-access-v8zfh") pod "e5697bae-d981-4bde-befa-a54c4cd0464b" (UID: "e5697bae-d981-4bde-befa-a54c4cd0464b"). InnerVolumeSpecName "kube-api-access-v8zfh". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:15:25 crc kubenswrapper[4967]: I1011 04:15:25.835966 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e5697bae-d981-4bde-befa-a54c4cd0464b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e5697bae-d981-4bde-befa-a54c4cd0464b" (UID: "e5697bae-d981-4bde-befa-a54c4cd0464b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:15:25 crc kubenswrapper[4967]: I1011 04:15:25.866384 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e5697bae-d981-4bde-befa-a54c4cd0464b-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 04:15:25 crc kubenswrapper[4967]: I1011 04:15:25.866702 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e5697bae-d981-4bde-befa-a54c4cd0464b-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 04:15:25 crc kubenswrapper[4967]: I1011 04:15:25.866877 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v8zfh\" (UniqueName: \"kubernetes.io/projected/e5697bae-d981-4bde-befa-a54c4cd0464b-kube-api-access-v8zfh\") on node \"crc\" DevicePath \"\"" Oct 11 04:15:26 crc kubenswrapper[4967]: I1011 04:15:26.213925 4967 generic.go:334] "Generic (PLEG): container finished" podID="e5697bae-d981-4bde-befa-a54c4cd0464b" containerID="3f6251b9c53845caf036eb3aa102ef71a72e18250507b0122d9b5a8207d2ddd3" exitCode=0 Oct 11 04:15:26 crc kubenswrapper[4967]: I1011 04:15:26.213981 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mplb4" event={"ID":"e5697bae-d981-4bde-befa-a54c4cd0464b","Type":"ContainerDied","Data":"3f6251b9c53845caf036eb3aa102ef71a72e18250507b0122d9b5a8207d2ddd3"} Oct 11 04:15:26 crc kubenswrapper[4967]: I1011 04:15:26.214010 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-mplb4" event={"ID":"e5697bae-d981-4bde-befa-a54c4cd0464b","Type":"ContainerDied","Data":"7402c4bd4fb52d770ef5d527ea90e8dec0df585a4800af6c6bf0cbdca21232c6"} Oct 11 04:15:26 crc kubenswrapper[4967]: I1011 04:15:26.214037 4967 scope.go:117] "RemoveContainer" containerID="3f6251b9c53845caf036eb3aa102ef71a72e18250507b0122d9b5a8207d2ddd3" Oct 11 04:15:26 crc kubenswrapper[4967]: I1011 04:15:26.214246 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-mplb4" Oct 11 04:15:26 crc kubenswrapper[4967]: I1011 04:15:26.267717 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-mplb4"] Oct 11 04:15:26 crc kubenswrapper[4967]: I1011 04:15:26.268136 4967 scope.go:117] "RemoveContainer" containerID="76880dd9c7e69eb10daa9639bcab7d47c49e95ed0706be5991a5ccbcd87cf339" Oct 11 04:15:26 crc kubenswrapper[4967]: I1011 04:15:26.274892 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-mplb4"] Oct 11 04:15:26 crc kubenswrapper[4967]: I1011 04:15:26.299630 4967 scope.go:117] "RemoveContainer" containerID="48ce4db879abc3f783ae7ced47ce5b09f31e0e899d6fffa286e04f94273e9fd5" Oct 11 04:15:26 crc kubenswrapper[4967]: I1011 04:15:26.350274 4967 scope.go:117] "RemoveContainer" containerID="3f6251b9c53845caf036eb3aa102ef71a72e18250507b0122d9b5a8207d2ddd3" Oct 11 04:15:26 crc kubenswrapper[4967]: E1011 04:15:26.350960 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3f6251b9c53845caf036eb3aa102ef71a72e18250507b0122d9b5a8207d2ddd3\": container with ID starting with 3f6251b9c53845caf036eb3aa102ef71a72e18250507b0122d9b5a8207d2ddd3 not found: ID does not exist" containerID="3f6251b9c53845caf036eb3aa102ef71a72e18250507b0122d9b5a8207d2ddd3" Oct 11 04:15:26 crc kubenswrapper[4967]: I1011 04:15:26.350997 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3f6251b9c53845caf036eb3aa102ef71a72e18250507b0122d9b5a8207d2ddd3"} err="failed to get container status \"3f6251b9c53845caf036eb3aa102ef71a72e18250507b0122d9b5a8207d2ddd3\": rpc error: code = NotFound desc = could not find container \"3f6251b9c53845caf036eb3aa102ef71a72e18250507b0122d9b5a8207d2ddd3\": container with ID starting with 3f6251b9c53845caf036eb3aa102ef71a72e18250507b0122d9b5a8207d2ddd3 not found: ID does not exist" Oct 11 04:15:26 crc kubenswrapper[4967]: I1011 04:15:26.351024 4967 scope.go:117] "RemoveContainer" containerID="76880dd9c7e69eb10daa9639bcab7d47c49e95ed0706be5991a5ccbcd87cf339" Oct 11 04:15:26 crc kubenswrapper[4967]: E1011 04:15:26.351367 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"76880dd9c7e69eb10daa9639bcab7d47c49e95ed0706be5991a5ccbcd87cf339\": container with ID starting with 76880dd9c7e69eb10daa9639bcab7d47c49e95ed0706be5991a5ccbcd87cf339 not found: ID does not exist" containerID="76880dd9c7e69eb10daa9639bcab7d47c49e95ed0706be5991a5ccbcd87cf339" Oct 11 04:15:26 crc kubenswrapper[4967]: I1011 04:15:26.351397 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"76880dd9c7e69eb10daa9639bcab7d47c49e95ed0706be5991a5ccbcd87cf339"} err="failed to get container status \"76880dd9c7e69eb10daa9639bcab7d47c49e95ed0706be5991a5ccbcd87cf339\": rpc error: code = NotFound desc = could not find container \"76880dd9c7e69eb10daa9639bcab7d47c49e95ed0706be5991a5ccbcd87cf339\": container with ID starting with 76880dd9c7e69eb10daa9639bcab7d47c49e95ed0706be5991a5ccbcd87cf339 not found: ID does not exist" Oct 11 04:15:26 crc kubenswrapper[4967]: I1011 04:15:26.351415 4967 scope.go:117] "RemoveContainer" containerID="48ce4db879abc3f783ae7ced47ce5b09f31e0e899d6fffa286e04f94273e9fd5" Oct 11 04:15:26 crc kubenswrapper[4967]: E1011 04:15:26.351595 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"48ce4db879abc3f783ae7ced47ce5b09f31e0e899d6fffa286e04f94273e9fd5\": container with ID starting with 48ce4db879abc3f783ae7ced47ce5b09f31e0e899d6fffa286e04f94273e9fd5 not found: ID does not exist" containerID="48ce4db879abc3f783ae7ced47ce5b09f31e0e899d6fffa286e04f94273e9fd5" Oct 11 04:15:26 crc kubenswrapper[4967]: I1011 04:15:26.351620 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"48ce4db879abc3f783ae7ced47ce5b09f31e0e899d6fffa286e04f94273e9fd5"} err="failed to get container status \"48ce4db879abc3f783ae7ced47ce5b09f31e0e899d6fffa286e04f94273e9fd5\": rpc error: code = NotFound desc = could not find container \"48ce4db879abc3f783ae7ced47ce5b09f31e0e899d6fffa286e04f94273e9fd5\": container with ID starting with 48ce4db879abc3f783ae7ced47ce5b09f31e0e899d6fffa286e04f94273e9fd5 not found: ID does not exist" Oct 11 04:15:26 crc kubenswrapper[4967]: I1011 04:15:26.833053 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e5697bae-d981-4bde-befa-a54c4cd0464b" path="/var/lib/kubelet/pods/e5697bae-d981-4bde-befa-a54c4cd0464b/volumes" Oct 11 04:15:44 crc kubenswrapper[4967]: I1011 04:15:44.089489 4967 patch_prober.go:28] interesting pod/machine-config-daemon-sgrv2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:15:44 crc kubenswrapper[4967]: I1011 04:15:44.090276 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:16:14 crc kubenswrapper[4967]: I1011 04:16:14.089287 4967 patch_prober.go:28] interesting pod/machine-config-daemon-sgrv2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:16:14 crc kubenswrapper[4967]: I1011 04:16:14.089849 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:16:14 crc kubenswrapper[4967]: I1011 04:16:14.089895 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" Oct 11 04:16:14 crc kubenswrapper[4967]: I1011 04:16:14.090640 4967 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"6010a3358387442fdc0ff1dfa8bf2d3fb285ef11cfb5f18ad38fe86548ca1f20"} pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 11 04:16:14 crc kubenswrapper[4967]: I1011 04:16:14.090685 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" containerID="cri-o://6010a3358387442fdc0ff1dfa8bf2d3fb285ef11cfb5f18ad38fe86548ca1f20" gracePeriod=600 Oct 11 04:16:14 crc kubenswrapper[4967]: E1011 04:16:14.213918 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:16:14 crc kubenswrapper[4967]: I1011 04:16:14.687509 4967 generic.go:334] "Generic (PLEG): container finished" podID="97c07678-14be-410c-b61f-498cb49bc960" containerID="6010a3358387442fdc0ff1dfa8bf2d3fb285ef11cfb5f18ad38fe86548ca1f20" exitCode=0 Oct 11 04:16:14 crc kubenswrapper[4967]: I1011 04:16:14.687547 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" event={"ID":"97c07678-14be-410c-b61f-498cb49bc960","Type":"ContainerDied","Data":"6010a3358387442fdc0ff1dfa8bf2d3fb285ef11cfb5f18ad38fe86548ca1f20"} Oct 11 04:16:14 crc kubenswrapper[4967]: I1011 04:16:14.687812 4967 scope.go:117] "RemoveContainer" containerID="c14dde3bbb545266e69254f79fceffdb0e7aa40ff2dad2c747eae9612aaecda9" Oct 11 04:16:14 crc kubenswrapper[4967]: I1011 04:16:14.688508 4967 scope.go:117] "RemoveContainer" containerID="6010a3358387442fdc0ff1dfa8bf2d3fb285ef11cfb5f18ad38fe86548ca1f20" Oct 11 04:16:14 crc kubenswrapper[4967]: E1011 04:16:14.688896 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:16:26 crc kubenswrapper[4967]: I1011 04:16:26.816899 4967 scope.go:117] "RemoveContainer" containerID="6010a3358387442fdc0ff1dfa8bf2d3fb285ef11cfb5f18ad38fe86548ca1f20" Oct 11 04:16:26 crc kubenswrapper[4967]: E1011 04:16:26.817999 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:16:37 crc kubenswrapper[4967]: I1011 04:16:37.815272 4967 scope.go:117] "RemoveContainer" containerID="6010a3358387442fdc0ff1dfa8bf2d3fb285ef11cfb5f18ad38fe86548ca1f20" Oct 11 04:16:37 crc kubenswrapper[4967]: E1011 04:16:37.816340 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:16:38 crc kubenswrapper[4967]: I1011 04:16:38.055162 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-create-nfr2c"] Oct 11 04:16:38 crc kubenswrapper[4967]: I1011 04:16:38.072597 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-create-zpqdx"] Oct 11 04:16:38 crc kubenswrapper[4967]: I1011 04:16:38.086510 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-create-zpqdx"] Oct 11 04:16:38 crc kubenswrapper[4967]: I1011 04:16:38.093558 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-create-nfr2c"] Oct 11 04:16:38 crc kubenswrapper[4967]: I1011 04:16:38.830380 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f395fae-7337-408a-ae83-1db82f8dc2b3" path="/var/lib/kubelet/pods/8f395fae-7337-408a-ae83-1db82f8dc2b3/volumes" Oct 11 04:16:38 crc kubenswrapper[4967]: I1011 04:16:38.834958 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="937176d5-2a41-4b55-89a8-4e2b0019b773" path="/var/lib/kubelet/pods/937176d5-2a41-4b55-89a8-4e2b0019b773/volumes" Oct 11 04:16:50 crc kubenswrapper[4967]: I1011 04:16:50.832292 4967 scope.go:117] "RemoveContainer" containerID="6010a3358387442fdc0ff1dfa8bf2d3fb285ef11cfb5f18ad38fe86548ca1f20" Oct 11 04:16:50 crc kubenswrapper[4967]: E1011 04:16:50.833115 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:16:51 crc kubenswrapper[4967]: I1011 04:16:51.030093 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-create-jrjv4"] Oct 11 04:16:51 crc kubenswrapper[4967]: I1011 04:16:51.039367 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-87d8-account-create-9sfk2"] Oct 11 04:16:51 crc kubenswrapper[4967]: I1011 04:16:51.048641 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-d5f2-account-create-dfps4"] Oct 11 04:16:51 crc kubenswrapper[4967]: I1011 04:16:51.057904 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-create-jrjv4"] Oct 11 04:16:51 crc kubenswrapper[4967]: I1011 04:16:51.066781 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-d5f2-account-create-dfps4"] Oct 11 04:16:51 crc kubenswrapper[4967]: I1011 04:16:51.076461 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-87d8-account-create-9sfk2"] Oct 11 04:16:52 crc kubenswrapper[4967]: I1011 04:16:52.849733 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="292067b2-43e9-46a1-8940-08bca18ce4f0" path="/var/lib/kubelet/pods/292067b2-43e9-46a1-8940-08bca18ce4f0/volumes" Oct 11 04:16:52 crc kubenswrapper[4967]: I1011 04:16:52.853423 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="35ca69d7-b265-4ab4-a047-5fbdb349f13d" path="/var/lib/kubelet/pods/35ca69d7-b265-4ab4-a047-5fbdb349f13d/volumes" Oct 11 04:16:52 crc kubenswrapper[4967]: I1011 04:16:52.855883 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b21fd878-e30e-498f-88f3-8a4f278ee0ec" path="/var/lib/kubelet/pods/b21fd878-e30e-498f-88f3-8a4f278ee0ec/volumes" Oct 11 04:17:03 crc kubenswrapper[4967]: I1011 04:17:03.051648 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-7892-account-create-2qg5w"] Oct 11 04:17:03 crc kubenswrapper[4967]: I1011 04:17:03.062169 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-7892-account-create-2qg5w"] Oct 11 04:17:04 crc kubenswrapper[4967]: I1011 04:17:04.833277 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3af91e9f-81cd-4cdb-89d2-a1c50f55f62a" path="/var/lib/kubelet/pods/3af91e9f-81cd-4cdb-89d2-a1c50f55f62a/volumes" Oct 11 04:17:05 crc kubenswrapper[4967]: I1011 04:17:05.815002 4967 scope.go:117] "RemoveContainer" containerID="6010a3358387442fdc0ff1dfa8bf2d3fb285ef11cfb5f18ad38fe86548ca1f20" Oct 11 04:17:05 crc kubenswrapper[4967]: E1011 04:17:05.815571 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:17:18 crc kubenswrapper[4967]: I1011 04:17:18.815313 4967 scope.go:117] "RemoveContainer" containerID="6010a3358387442fdc0ff1dfa8bf2d3fb285ef11cfb5f18ad38fe86548ca1f20" Oct 11 04:17:18 crc kubenswrapper[4967]: E1011 04:17:18.816095 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:17:20 crc kubenswrapper[4967]: I1011 04:17:20.032626 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-create-f5bhd"] Oct 11 04:17:20 crc kubenswrapper[4967]: I1011 04:17:20.043129 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-create-f5bhd"] Oct 11 04:17:20 crc kubenswrapper[4967]: I1011 04:17:20.827299 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6f16161f-b863-46ca-b2f1-7a6248f27c4d" path="/var/lib/kubelet/pods/6f16161f-b863-46ca-b2f1-7a6248f27c4d/volumes" Oct 11 04:17:23 crc kubenswrapper[4967]: I1011 04:17:23.034320 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-create-9p6rn"] Oct 11 04:17:23 crc kubenswrapper[4967]: I1011 04:17:23.047303 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-create-257lm"] Oct 11 04:17:23 crc kubenswrapper[4967]: I1011 04:17:23.054945 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-create-9p6rn"] Oct 11 04:17:23 crc kubenswrapper[4967]: I1011 04:17:23.061448 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-create-257lm"] Oct 11 04:17:23 crc kubenswrapper[4967]: I1011 04:17:23.290499 4967 scope.go:117] "RemoveContainer" containerID="99d6af2f22ee5cb47dba8ad27385fb8bd4ee168973097919bb99a148169b6a28" Oct 11 04:17:23 crc kubenswrapper[4967]: I1011 04:17:23.363611 4967 scope.go:117] "RemoveContainer" containerID="1ae031598cabc6c10cadd267df2488dad0c87a931f779e39bab402b4c50ce80b" Oct 11 04:17:23 crc kubenswrapper[4967]: I1011 04:17:23.401964 4967 scope.go:117] "RemoveContainer" containerID="5db53328505e2416c3f4d98b1042bc99fe763631c2946f9777368c2f0241df8c" Oct 11 04:17:23 crc kubenswrapper[4967]: I1011 04:17:23.442303 4967 scope.go:117] "RemoveContainer" containerID="81fd1c2ab306fc3598137e0222bb8f156cbf18904c6b0d8cf76277de06301ba5" Oct 11 04:17:23 crc kubenswrapper[4967]: I1011 04:17:23.482999 4967 scope.go:117] "RemoveContainer" containerID="0fcbd3c9399c32b7d84f732ff8c9b109eb4a2f153f4c9cd8cc2852bb1f1e5fbe" Oct 11 04:17:23 crc kubenswrapper[4967]: I1011 04:17:23.522649 4967 scope.go:117] "RemoveContainer" containerID="a29180c1a7204d0a126ff43042e2ee8b3dd7262ae170a94a670169d29b5a7fd7" Oct 11 04:17:23 crc kubenswrapper[4967]: I1011 04:17:23.566285 4967 scope.go:117] "RemoveContainer" containerID="2a41c7134fd43de5cbd967564c77fb546d537bef1bb8b9b99d570a3b67c21e9c" Oct 11 04:17:24 crc kubenswrapper[4967]: I1011 04:17:24.831032 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d9c71007-45ce-4684-87eb-1884ec5192c9" path="/var/lib/kubelet/pods/d9c71007-45ce-4684-87eb-1884ec5192c9/volumes" Oct 11 04:17:24 crc kubenswrapper[4967]: I1011 04:17:24.832029 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fdb459dd-26c4-49c4-bf1f-dad807314e3d" path="/var/lib/kubelet/pods/fdb459dd-26c4-49c4-bf1f-dad807314e3d/volumes" Oct 11 04:17:26 crc kubenswrapper[4967]: I1011 04:17:26.030205 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/glance-db-sync-mgcb9"] Oct 11 04:17:26 crc kubenswrapper[4967]: I1011 04:17:26.042397 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/glance-db-sync-mgcb9"] Oct 11 04:17:26 crc kubenswrapper[4967]: I1011 04:17:26.832741 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="df08696a-e0e9-4d45-9ffb-c41eff1b5fb4" path="/var/lib/kubelet/pods/df08696a-e0e9-4d45-9ffb-c41eff1b5fb4/volumes" Oct 11 04:17:27 crc kubenswrapper[4967]: I1011 04:17:27.033050 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-db-sync-ks59l"] Oct 11 04:17:27 crc kubenswrapper[4967]: I1011 04:17:27.050767 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-db-sync-ks59l"] Oct 11 04:17:28 crc kubenswrapper[4967]: I1011 04:17:28.826880 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0ccef7a0-eebe-4794-b38d-690a98ee2055" path="/var/lib/kubelet/pods/0ccef7a0-eebe-4794-b38d-690a98ee2055/volumes" Oct 11 04:17:29 crc kubenswrapper[4967]: I1011 04:17:29.814991 4967 scope.go:117] "RemoveContainer" containerID="6010a3358387442fdc0ff1dfa8bf2d3fb285ef11cfb5f18ad38fe86548ca1f20" Oct 11 04:17:29 crc kubenswrapper[4967]: E1011 04:17:29.815313 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:17:32 crc kubenswrapper[4967]: I1011 04:17:32.064863 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-db-create-4m9mw"] Oct 11 04:17:32 crc kubenswrapper[4967]: I1011 04:17:32.073934 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ironic-db-create-4m9mw"] Oct 11 04:17:32 crc kubenswrapper[4967]: I1011 04:17:32.828229 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c70d7931-0ef5-44dc-88dd-b0f7b5ef3afd" path="/var/lib/kubelet/pods/c70d7931-0ef5-44dc-88dd-b0f7b5ef3afd/volumes" Oct 11 04:17:40 crc kubenswrapper[4967]: I1011 04:17:40.028647 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-5b20-account-create-lb22f"] Oct 11 04:17:40 crc kubenswrapper[4967]: I1011 04:17:40.038916 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-317e-account-create-l88bq"] Oct 11 04:17:40 crc kubenswrapper[4967]: I1011 04:17:40.047745 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-b879-account-create-6v2r9"] Oct 11 04:17:40 crc kubenswrapper[4967]: I1011 04:17:40.055552 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-5b20-account-create-lb22f"] Oct 11 04:17:40 crc kubenswrapper[4967]: I1011 04:17:40.066408 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-b879-account-create-6v2r9"] Oct 11 04:17:40 crc kubenswrapper[4967]: I1011 04:17:40.075817 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-317e-account-create-l88bq"] Oct 11 04:17:40 crc kubenswrapper[4967]: I1011 04:17:40.844816 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2743338e-1478-4402-b29f-d4b86722d241" path="/var/lib/kubelet/pods/2743338e-1478-4402-b29f-d4b86722d241/volumes" Oct 11 04:17:40 crc kubenswrapper[4967]: I1011 04:17:40.846743 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2930d6b7-0aad-4123-b8bd-4dad4b8768fe" path="/var/lib/kubelet/pods/2930d6b7-0aad-4123-b8bd-4dad4b8768fe/volumes" Oct 11 04:17:40 crc kubenswrapper[4967]: I1011 04:17:40.847436 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="92ccb4c4-e0b5-4592-bdfc-9909ee48b8b1" path="/var/lib/kubelet/pods/92ccb4c4-e0b5-4592-bdfc-9909ee48b8b1/volumes" Oct 11 04:17:41 crc kubenswrapper[4967]: I1011 04:17:41.814929 4967 scope.go:117] "RemoveContainer" containerID="6010a3358387442fdc0ff1dfa8bf2d3fb285ef11cfb5f18ad38fe86548ca1f20" Oct 11 04:17:41 crc kubenswrapper[4967]: E1011 04:17:41.815702 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:17:47 crc kubenswrapper[4967]: I1011 04:17:47.044444 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-8389-account-create-pvbnb"] Oct 11 04:17:47 crc kubenswrapper[4967]: I1011 04:17:47.063056 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ironic-8389-account-create-pvbnb"] Oct 11 04:17:47 crc kubenswrapper[4967]: I1011 04:17:47.072953 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/placement-db-sync-zf6xs"] Oct 11 04:17:47 crc kubenswrapper[4967]: I1011 04:17:47.080623 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/keystone-bootstrap-5dtlx"] Oct 11 04:17:47 crc kubenswrapper[4967]: I1011 04:17:47.087138 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/keystone-bootstrap-5dtlx"] Oct 11 04:17:47 crc kubenswrapper[4967]: I1011 04:17:47.093418 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/placement-db-sync-zf6xs"] Oct 11 04:17:48 crc kubenswrapper[4967]: I1011 04:17:48.827857 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="05059c3c-e29a-4bfa-a0d4-0d56b3d85af9" path="/var/lib/kubelet/pods/05059c3c-e29a-4bfa-a0d4-0d56b3d85af9/volumes" Oct 11 04:17:48 crc kubenswrapper[4967]: I1011 04:17:48.828747 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4691fb42-e6bb-4e1b-a4c2-1f65e1292d4d" path="/var/lib/kubelet/pods/4691fb42-e6bb-4e1b-a4c2-1f65e1292d4d/volumes" Oct 11 04:17:48 crc kubenswrapper[4967]: I1011 04:17:48.829522 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3" path="/var/lib/kubelet/pods/dcea46f6-5b3f-42fe-b1f7-f9a66582d8e3/volumes" Oct 11 04:17:52 crc kubenswrapper[4967]: I1011 04:17:52.828709 4967 scope.go:117] "RemoveContainer" containerID="6010a3358387442fdc0ff1dfa8bf2d3fb285ef11cfb5f18ad38fe86548ca1f20" Oct 11 04:17:52 crc kubenswrapper[4967]: E1011 04:17:52.830024 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:18:05 crc kubenswrapper[4967]: I1011 04:18:05.814992 4967 scope.go:117] "RemoveContainer" containerID="6010a3358387442fdc0ff1dfa8bf2d3fb285ef11cfb5f18ad38fe86548ca1f20" Oct 11 04:18:05 crc kubenswrapper[4967]: E1011 04:18:05.816467 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:18:14 crc kubenswrapper[4967]: I1011 04:18:14.252195 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-vt6vn/must-gather-689h5"] Oct 11 04:18:14 crc kubenswrapper[4967]: E1011 04:18:14.253310 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5697bae-d981-4bde-befa-a54c4cd0464b" containerName="extract-utilities" Oct 11 04:18:14 crc kubenswrapper[4967]: I1011 04:18:14.253359 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5697bae-d981-4bde-befa-a54c4cd0464b" containerName="extract-utilities" Oct 11 04:18:14 crc kubenswrapper[4967]: E1011 04:18:14.253402 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a16f74ee-c779-4fd9-b945-3e455927298a" containerName="collect-profiles" Oct 11 04:18:14 crc kubenswrapper[4967]: I1011 04:18:14.253413 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="a16f74ee-c779-4fd9-b945-3e455927298a" containerName="collect-profiles" Oct 11 04:18:14 crc kubenswrapper[4967]: E1011 04:18:14.253432 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5697bae-d981-4bde-befa-a54c4cd0464b" containerName="extract-content" Oct 11 04:18:14 crc kubenswrapper[4967]: I1011 04:18:14.253444 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5697bae-d981-4bde-befa-a54c4cd0464b" containerName="extract-content" Oct 11 04:18:14 crc kubenswrapper[4967]: E1011 04:18:14.253468 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e5697bae-d981-4bde-befa-a54c4cd0464b" containerName="registry-server" Oct 11 04:18:14 crc kubenswrapper[4967]: I1011 04:18:14.253478 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="e5697bae-d981-4bde-befa-a54c4cd0464b" containerName="registry-server" Oct 11 04:18:14 crc kubenswrapper[4967]: I1011 04:18:14.253726 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="e5697bae-d981-4bde-befa-a54c4cd0464b" containerName="registry-server" Oct 11 04:18:14 crc kubenswrapper[4967]: I1011 04:18:14.253751 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="a16f74ee-c779-4fd9-b945-3e455927298a" containerName="collect-profiles" Oct 11 04:18:14 crc kubenswrapper[4967]: I1011 04:18:14.255242 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vt6vn/must-gather-689h5" Oct 11 04:18:14 crc kubenswrapper[4967]: I1011 04:18:14.265420 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-vt6vn/must-gather-689h5"] Oct 11 04:18:14 crc kubenswrapper[4967]: I1011 04:18:14.269649 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-vt6vn"/"default-dockercfg-9jjh7" Oct 11 04:18:14 crc kubenswrapper[4967]: I1011 04:18:14.269937 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-vt6vn"/"openshift-service-ca.crt" Oct 11 04:18:14 crc kubenswrapper[4967]: I1011 04:18:14.270118 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-vt6vn"/"kube-root-ca.crt" Oct 11 04:18:14 crc kubenswrapper[4967]: I1011 04:18:14.407312 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/31a11147-0dcb-4957-826d-6579699bde94-must-gather-output\") pod \"must-gather-689h5\" (UID: \"31a11147-0dcb-4957-826d-6579699bde94\") " pod="openshift-must-gather-vt6vn/must-gather-689h5" Oct 11 04:18:14 crc kubenswrapper[4967]: I1011 04:18:14.407576 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hxxgw\" (UniqueName: \"kubernetes.io/projected/31a11147-0dcb-4957-826d-6579699bde94-kube-api-access-hxxgw\") pod \"must-gather-689h5\" (UID: \"31a11147-0dcb-4957-826d-6579699bde94\") " pod="openshift-must-gather-vt6vn/must-gather-689h5" Oct 11 04:18:14 crc kubenswrapper[4967]: I1011 04:18:14.509691 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/31a11147-0dcb-4957-826d-6579699bde94-must-gather-output\") pod \"must-gather-689h5\" (UID: \"31a11147-0dcb-4957-826d-6579699bde94\") " pod="openshift-must-gather-vt6vn/must-gather-689h5" Oct 11 04:18:14 crc kubenswrapper[4967]: I1011 04:18:14.509804 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hxxgw\" (UniqueName: \"kubernetes.io/projected/31a11147-0dcb-4957-826d-6579699bde94-kube-api-access-hxxgw\") pod \"must-gather-689h5\" (UID: \"31a11147-0dcb-4957-826d-6579699bde94\") " pod="openshift-must-gather-vt6vn/must-gather-689h5" Oct 11 04:18:14 crc kubenswrapper[4967]: I1011 04:18:14.510188 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/31a11147-0dcb-4957-826d-6579699bde94-must-gather-output\") pod \"must-gather-689h5\" (UID: \"31a11147-0dcb-4957-826d-6579699bde94\") " pod="openshift-must-gather-vt6vn/must-gather-689h5" Oct 11 04:18:14 crc kubenswrapper[4967]: I1011 04:18:14.530264 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hxxgw\" (UniqueName: \"kubernetes.io/projected/31a11147-0dcb-4957-826d-6579699bde94-kube-api-access-hxxgw\") pod \"must-gather-689h5\" (UID: \"31a11147-0dcb-4957-826d-6579699bde94\") " pod="openshift-must-gather-vt6vn/must-gather-689h5" Oct 11 04:18:14 crc kubenswrapper[4967]: I1011 04:18:14.588325 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vt6vn/must-gather-689h5" Oct 11 04:18:15 crc kubenswrapper[4967]: I1011 04:18:15.089559 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-vt6vn/must-gather-689h5"] Oct 11 04:18:16 crc kubenswrapper[4967]: I1011 04:18:16.042044 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/barbican-db-sync-wdc9t"] Oct 11 04:18:16 crc kubenswrapper[4967]: I1011 04:18:16.051396 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/barbican-db-sync-wdc9t"] Oct 11 04:18:16 crc kubenswrapper[4967]: I1011 04:18:16.055066 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vt6vn/must-gather-689h5" event={"ID":"31a11147-0dcb-4957-826d-6579699bde94","Type":"ContainerStarted","Data":"89986b60de81611ef8611113a78377f0120bc7a756fc18abde5d138cb27c7884"} Oct 11 04:18:16 crc kubenswrapper[4967]: I1011 04:18:16.826393 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8539e0a2-fecc-46f8-aa2f-a931a31cf17f" path="/var/lib/kubelet/pods/8539e0a2-fecc-46f8-aa2f-a931a31cf17f/volumes" Oct 11 04:18:17 crc kubenswrapper[4967]: I1011 04:18:17.032018 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/neutron-db-sync-nqrkv"] Oct 11 04:18:17 crc kubenswrapper[4967]: I1011 04:18:17.045659 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/neutron-db-sync-nqrkv"] Oct 11 04:18:18 crc kubenswrapper[4967]: I1011 04:18:18.829464 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="28309ac7-8bd7-425c-a2f8-1041f1749acb" path="/var/lib/kubelet/pods/28309ac7-8bd7-425c-a2f8-1041f1749acb/volumes" Oct 11 04:18:19 crc kubenswrapper[4967]: I1011 04:18:19.815565 4967 scope.go:117] "RemoveContainer" containerID="6010a3358387442fdc0ff1dfa8bf2d3fb285ef11cfb5f18ad38fe86548ca1f20" Oct 11 04:18:19 crc kubenswrapper[4967]: E1011 04:18:19.815942 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:18:22 crc kubenswrapper[4967]: I1011 04:18:22.120022 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vt6vn/must-gather-689h5" event={"ID":"31a11147-0dcb-4957-826d-6579699bde94","Type":"ContainerStarted","Data":"f672d23f2106ef666680d24be5d2207af3059c78ee3b47c6c60c33f34e34e223"} Oct 11 04:18:23 crc kubenswrapper[4967]: I1011 04:18:23.131994 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vt6vn/must-gather-689h5" event={"ID":"31a11147-0dcb-4957-826d-6579699bde94","Type":"ContainerStarted","Data":"f6b72a8eac803d353e976a61f4ebfb8f5c2e6ad41329bb8a50f24479247338a7"} Oct 11 04:18:23 crc kubenswrapper[4967]: I1011 04:18:23.152059 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-vt6vn/must-gather-689h5" podStartSLOduration=2.5128189130000003 podStartE2EDuration="9.152034954s" podCreationTimestamp="2025-10-11 04:18:14 +0000 UTC" firstStartedPulling="2025-10-11 04:18:15.08324395 +0000 UTC m=+1623.046452883" lastFinishedPulling="2025-10-11 04:18:21.722459971 +0000 UTC m=+1629.685668924" observedRunningTime="2025-10-11 04:18:23.151901791 +0000 UTC m=+1631.115110744" watchObservedRunningTime="2025-10-11 04:18:23.152034954 +0000 UTC m=+1631.115243897" Oct 11 04:18:23 crc kubenswrapper[4967]: I1011 04:18:23.758637 4967 scope.go:117] "RemoveContainer" containerID="dc76d506020f9eaa727c4fa0b31d77be06538ced515178c9fdefb0d2a99bb700" Oct 11 04:18:23 crc kubenswrapper[4967]: I1011 04:18:23.795692 4967 scope.go:117] "RemoveContainer" containerID="d0751b3ff58d0130b73e6aa500956566e743d151ef828a30c0b59e0ae23c9a49" Oct 11 04:18:23 crc kubenswrapper[4967]: I1011 04:18:23.849031 4967 scope.go:117] "RemoveContainer" containerID="6fb4470936d71c0ec45dcea895c56f029bbb9d071929b0c549bb74cb2065a671" Oct 11 04:18:23 crc kubenswrapper[4967]: I1011 04:18:23.922252 4967 scope.go:117] "RemoveContainer" containerID="4b56fc23eeae7214925b7a9914c1b342181d9803ea7a698b33723c05b84b8292" Oct 11 04:18:23 crc kubenswrapper[4967]: I1011 04:18:23.991023 4967 scope.go:117] "RemoveContainer" containerID="4f4319ea2749661e591abb39022c4bacf81f997235ba5a04b06afb53e61abb94" Oct 11 04:18:24 crc kubenswrapper[4967]: I1011 04:18:24.039420 4967 scope.go:117] "RemoveContainer" containerID="208f95e3e45952d983d55b0410f07c0a8a43bff9ada9afb6bcc1da3bf5f1f71e" Oct 11 04:18:24 crc kubenswrapper[4967]: I1011 04:18:24.062618 4967 scope.go:117] "RemoveContainer" containerID="d63c539df8be1bba90129b4002c49f274647e7342820081417c3303fbbfaead7" Oct 11 04:18:24 crc kubenswrapper[4967]: I1011 04:18:24.098034 4967 scope.go:117] "RemoveContainer" containerID="abdc71002731f1f79b9c3e24a5bce139cf8e8ef69665b9d7a89453b7eda936ce" Oct 11 04:18:24 crc kubenswrapper[4967]: I1011 04:18:24.125758 4967 scope.go:117] "RemoveContainer" containerID="05e2a28172819cddfb09e990e95790043afc1033045364b6fd724ba8b52a2b0b" Oct 11 04:18:24 crc kubenswrapper[4967]: I1011 04:18:24.164313 4967 scope.go:117] "RemoveContainer" containerID="3a312f975d52e911748a61f193ab7ac5430ad2106363aacfd23680c68c6d79a9" Oct 11 04:18:24 crc kubenswrapper[4967]: I1011 04:18:24.185019 4967 scope.go:117] "RemoveContainer" containerID="e76735772140c83529136bb7bb0ee17f72e4d64e36fa332ded89a7ba6254197b" Oct 11 04:18:24 crc kubenswrapper[4967]: I1011 04:18:24.220860 4967 scope.go:117] "RemoveContainer" containerID="31a36b28144237ec04957225c745f8fc483b4ed7206adce1567814f143f87402" Oct 11 04:18:24 crc kubenswrapper[4967]: I1011 04:18:24.242172 4967 scope.go:117] "RemoveContainer" containerID="240ba3af22688386cb63250f6b108ce87c2b90ac426742afa8ae2ed6c30da0cd" Oct 11 04:18:24 crc kubenswrapper[4967]: E1011 04:18:24.745656 4967 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.50:42530->38.102.83.50:40947: write tcp 38.102.83.50:42530->38.102.83.50:40947: write: broken pipe Oct 11 04:18:25 crc kubenswrapper[4967]: I1011 04:18:25.435598 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-vt6vn/crc-debug-n46l6"] Oct 11 04:18:25 crc kubenswrapper[4967]: I1011 04:18:25.437006 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vt6vn/crc-debug-n46l6" Oct 11 04:18:25 crc kubenswrapper[4967]: I1011 04:18:25.561757 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h5hld\" (UniqueName: \"kubernetes.io/projected/b9d03cd0-766f-4fde-a0bc-95693e4ae111-kube-api-access-h5hld\") pod \"crc-debug-n46l6\" (UID: \"b9d03cd0-766f-4fde-a0bc-95693e4ae111\") " pod="openshift-must-gather-vt6vn/crc-debug-n46l6" Oct 11 04:18:25 crc kubenswrapper[4967]: I1011 04:18:25.561823 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9d03cd0-766f-4fde-a0bc-95693e4ae111-host\") pod \"crc-debug-n46l6\" (UID: \"b9d03cd0-766f-4fde-a0bc-95693e4ae111\") " pod="openshift-must-gather-vt6vn/crc-debug-n46l6" Oct 11 04:18:25 crc kubenswrapper[4967]: I1011 04:18:25.663628 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h5hld\" (UniqueName: \"kubernetes.io/projected/b9d03cd0-766f-4fde-a0bc-95693e4ae111-kube-api-access-h5hld\") pod \"crc-debug-n46l6\" (UID: \"b9d03cd0-766f-4fde-a0bc-95693e4ae111\") " pod="openshift-must-gather-vt6vn/crc-debug-n46l6" Oct 11 04:18:25 crc kubenswrapper[4967]: I1011 04:18:25.663691 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9d03cd0-766f-4fde-a0bc-95693e4ae111-host\") pod \"crc-debug-n46l6\" (UID: \"b9d03cd0-766f-4fde-a0bc-95693e4ae111\") " pod="openshift-must-gather-vt6vn/crc-debug-n46l6" Oct 11 04:18:25 crc kubenswrapper[4967]: I1011 04:18:25.663794 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9d03cd0-766f-4fde-a0bc-95693e4ae111-host\") pod \"crc-debug-n46l6\" (UID: \"b9d03cd0-766f-4fde-a0bc-95693e4ae111\") " pod="openshift-must-gather-vt6vn/crc-debug-n46l6" Oct 11 04:18:25 crc kubenswrapper[4967]: I1011 04:18:25.689000 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h5hld\" (UniqueName: \"kubernetes.io/projected/b9d03cd0-766f-4fde-a0bc-95693e4ae111-kube-api-access-h5hld\") pod \"crc-debug-n46l6\" (UID: \"b9d03cd0-766f-4fde-a0bc-95693e4ae111\") " pod="openshift-must-gather-vt6vn/crc-debug-n46l6" Oct 11 04:18:25 crc kubenswrapper[4967]: I1011 04:18:25.754657 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vt6vn/crc-debug-n46l6" Oct 11 04:18:26 crc kubenswrapper[4967]: I1011 04:18:26.177642 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vt6vn/crc-debug-n46l6" event={"ID":"b9d03cd0-766f-4fde-a0bc-95693e4ae111","Type":"ContainerStarted","Data":"c9c3c8f2053aed4914b39a53642db1ff6539bab6315e40a5382f8b5c966c702f"} Oct 11 04:18:31 crc kubenswrapper[4967]: I1011 04:18:31.814927 4967 scope.go:117] "RemoveContainer" containerID="6010a3358387442fdc0ff1dfa8bf2d3fb285ef11cfb5f18ad38fe86548ca1f20" Oct 11 04:18:31 crc kubenswrapper[4967]: E1011 04:18:31.815570 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:18:33 crc kubenswrapper[4967]: I1011 04:18:33.040978 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/cinder-db-sync-6bhhf"] Oct 11 04:18:33 crc kubenswrapper[4967]: I1011 04:18:33.047819 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/cinder-db-sync-6bhhf"] Oct 11 04:18:34 crc kubenswrapper[4967]: I1011 04:18:34.826734 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1680354a-b333-4e21-9748-0062d78a479c" path="/var/lib/kubelet/pods/1680354a-b333-4e21-9748-0062d78a479c/volumes" Oct 11 04:18:40 crc kubenswrapper[4967]: I1011 04:18:40.307957 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vt6vn/crc-debug-n46l6" event={"ID":"b9d03cd0-766f-4fde-a0bc-95693e4ae111","Type":"ContainerStarted","Data":"7a838cf131945ab75399cf021780705d1d3db35b4babc83434cd6d5b26cdd25a"} Oct 11 04:18:40 crc kubenswrapper[4967]: I1011 04:18:40.327546 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-vt6vn/crc-debug-n46l6" podStartSLOduration=1.841307338 podStartE2EDuration="15.327530738s" podCreationTimestamp="2025-10-11 04:18:25 +0000 UTC" firstStartedPulling="2025-10-11 04:18:25.793839468 +0000 UTC m=+1633.757048401" lastFinishedPulling="2025-10-11 04:18:39.280062868 +0000 UTC m=+1647.243271801" observedRunningTime="2025-10-11 04:18:40.319102521 +0000 UTC m=+1648.282311454" watchObservedRunningTime="2025-10-11 04:18:40.327530738 +0000 UTC m=+1648.290739671" Oct 11 04:18:44 crc kubenswrapper[4967]: I1011 04:18:44.814810 4967 scope.go:117] "RemoveContainer" containerID="6010a3358387442fdc0ff1dfa8bf2d3fb285ef11cfb5f18ad38fe86548ca1f20" Oct 11 04:18:44 crc kubenswrapper[4967]: E1011 04:18:44.815711 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:18:48 crc kubenswrapper[4967]: I1011 04:18:48.029151 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-inspector-db-create-bprvt"] Oct 11 04:18:48 crc kubenswrapper[4967]: I1011 04:18:48.036538 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ironic-inspector-db-create-bprvt"] Oct 11 04:18:48 crc kubenswrapper[4967]: I1011 04:18:48.825337 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="12c05ddc-4b39-4924-924c-955b8639a839" path="/var/lib/kubelet/pods/12c05ddc-4b39-4924-924c-955b8639a839/volumes" Oct 11 04:18:57 crc kubenswrapper[4967]: I1011 04:18:57.030773 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-db-create-zt7cl"] Oct 11 04:18:57 crc kubenswrapper[4967]: I1011 04:18:57.041531 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-db-create-zt7cl"] Oct 11 04:18:57 crc kubenswrapper[4967]: I1011 04:18:57.815193 4967 scope.go:117] "RemoveContainer" containerID="6010a3358387442fdc0ff1dfa8bf2d3fb285ef11cfb5f18ad38fe86548ca1f20" Oct 11 04:18:57 crc kubenswrapper[4967]: E1011 04:18:57.815487 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:18:58 crc kubenswrapper[4967]: I1011 04:18:58.826727 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="540b1dfa-8e24-4aed-a3d9-6c2a1e367155" path="/var/lib/kubelet/pods/540b1dfa-8e24-4aed-a3d9-6c2a1e367155/volumes" Oct 11 04:18:59 crc kubenswrapper[4967]: I1011 04:18:59.043195 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-db-create-zb7zk"] Oct 11 04:18:59 crc kubenswrapper[4967]: I1011 04:18:59.052913 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-db-create-zth24"] Oct 11 04:18:59 crc kubenswrapper[4967]: I1011 04:18:59.061700 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-db-create-zb7zk"] Oct 11 04:18:59 crc kubenswrapper[4967]: I1011 04:18:59.068360 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/ironic-inspector-ba91-account-create-595xr"] Oct 11 04:18:59 crc kubenswrapper[4967]: I1011 04:18:59.075232 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-db-create-zth24"] Oct 11 04:18:59 crc kubenswrapper[4967]: I1011 04:18:59.083033 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/ironic-inspector-ba91-account-create-595xr"] Oct 11 04:18:59 crc kubenswrapper[4967]: I1011 04:18:59.791459 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-6pmqm"] Oct 11 04:18:59 crc kubenswrapper[4967]: I1011 04:18:59.794566 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6pmqm" Oct 11 04:18:59 crc kubenswrapper[4967]: I1011 04:18:59.804015 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6pmqm"] Oct 11 04:18:59 crc kubenswrapper[4967]: I1011 04:18:59.921322 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e2467e3-03a7-43d6-806b-0e4f304d7983-utilities\") pod \"certified-operators-6pmqm\" (UID: \"5e2467e3-03a7-43d6-806b-0e4f304d7983\") " pod="openshift-marketplace/certified-operators-6pmqm" Oct 11 04:18:59 crc kubenswrapper[4967]: I1011 04:18:59.921710 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cgjj4\" (UniqueName: \"kubernetes.io/projected/5e2467e3-03a7-43d6-806b-0e4f304d7983-kube-api-access-cgjj4\") pod \"certified-operators-6pmqm\" (UID: \"5e2467e3-03a7-43d6-806b-0e4f304d7983\") " pod="openshift-marketplace/certified-operators-6pmqm" Oct 11 04:18:59 crc kubenswrapper[4967]: I1011 04:18:59.921847 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e2467e3-03a7-43d6-806b-0e4f304d7983-catalog-content\") pod \"certified-operators-6pmqm\" (UID: \"5e2467e3-03a7-43d6-806b-0e4f304d7983\") " pod="openshift-marketplace/certified-operators-6pmqm" Oct 11 04:19:00 crc kubenswrapper[4967]: I1011 04:19:00.023604 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e2467e3-03a7-43d6-806b-0e4f304d7983-catalog-content\") pod \"certified-operators-6pmqm\" (UID: \"5e2467e3-03a7-43d6-806b-0e4f304d7983\") " pod="openshift-marketplace/certified-operators-6pmqm" Oct 11 04:19:00 crc kubenswrapper[4967]: I1011 04:19:00.023800 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e2467e3-03a7-43d6-806b-0e4f304d7983-utilities\") pod \"certified-operators-6pmqm\" (UID: \"5e2467e3-03a7-43d6-806b-0e4f304d7983\") " pod="openshift-marketplace/certified-operators-6pmqm" Oct 11 04:19:00 crc kubenswrapper[4967]: I1011 04:19:00.023836 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cgjj4\" (UniqueName: \"kubernetes.io/projected/5e2467e3-03a7-43d6-806b-0e4f304d7983-kube-api-access-cgjj4\") pod \"certified-operators-6pmqm\" (UID: \"5e2467e3-03a7-43d6-806b-0e4f304d7983\") " pod="openshift-marketplace/certified-operators-6pmqm" Oct 11 04:19:00 crc kubenswrapper[4967]: I1011 04:19:00.024391 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e2467e3-03a7-43d6-806b-0e4f304d7983-catalog-content\") pod \"certified-operators-6pmqm\" (UID: \"5e2467e3-03a7-43d6-806b-0e4f304d7983\") " pod="openshift-marketplace/certified-operators-6pmqm" Oct 11 04:19:00 crc kubenswrapper[4967]: I1011 04:19:00.024416 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e2467e3-03a7-43d6-806b-0e4f304d7983-utilities\") pod \"certified-operators-6pmqm\" (UID: \"5e2467e3-03a7-43d6-806b-0e4f304d7983\") " pod="openshift-marketplace/certified-operators-6pmqm" Oct 11 04:19:00 crc kubenswrapper[4967]: I1011 04:19:00.049012 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cgjj4\" (UniqueName: \"kubernetes.io/projected/5e2467e3-03a7-43d6-806b-0e4f304d7983-kube-api-access-cgjj4\") pod \"certified-operators-6pmqm\" (UID: \"5e2467e3-03a7-43d6-806b-0e4f304d7983\") " pod="openshift-marketplace/certified-operators-6pmqm" Oct 11 04:19:00 crc kubenswrapper[4967]: I1011 04:19:00.116035 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6pmqm" Oct 11 04:19:00 crc kubenswrapper[4967]: I1011 04:19:00.615683 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-6pmqm"] Oct 11 04:19:00 crc kubenswrapper[4967]: I1011 04:19:00.825777 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="098dc82c-7d36-48e4-bc48-28e8721f6617" path="/var/lib/kubelet/pods/098dc82c-7d36-48e4-bc48-28e8721f6617/volumes" Oct 11 04:19:00 crc kubenswrapper[4967]: I1011 04:19:00.826737 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="706a95a7-728b-40e8-8c4e-221cc836295f" path="/var/lib/kubelet/pods/706a95a7-728b-40e8-8c4e-221cc836295f/volumes" Oct 11 04:19:00 crc kubenswrapper[4967]: I1011 04:19:00.827214 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="de9db0f0-79dd-4edc-8022-c32a3a1340a1" path="/var/lib/kubelet/pods/de9db0f0-79dd-4edc-8022-c32a3a1340a1/volumes" Oct 11 04:19:01 crc kubenswrapper[4967]: I1011 04:19:01.547663 4967 generic.go:334] "Generic (PLEG): container finished" podID="5e2467e3-03a7-43d6-806b-0e4f304d7983" containerID="3ae5667b452159e6da15dc765036580d6cf5a9bb54f533f07fe0be953390f943" exitCode=0 Oct 11 04:19:01 crc kubenswrapper[4967]: I1011 04:19:01.547721 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6pmqm" event={"ID":"5e2467e3-03a7-43d6-806b-0e4f304d7983","Type":"ContainerDied","Data":"3ae5667b452159e6da15dc765036580d6cf5a9bb54f533f07fe0be953390f943"} Oct 11 04:19:01 crc kubenswrapper[4967]: I1011 04:19:01.547753 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6pmqm" event={"ID":"5e2467e3-03a7-43d6-806b-0e4f304d7983","Type":"ContainerStarted","Data":"715b02e7e50cf7a922f2674d946f8452d05d0ef556bbfe8a182c9596e532f0ef"} Oct 11 04:19:03 crc kubenswrapper[4967]: I1011 04:19:03.567847 4967 generic.go:334] "Generic (PLEG): container finished" podID="5e2467e3-03a7-43d6-806b-0e4f304d7983" containerID="aa89d19790253b752a83e82914e79b50993bb43bfe09cc7f656a09f6c250516d" exitCode=0 Oct 11 04:19:03 crc kubenswrapper[4967]: I1011 04:19:03.567974 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6pmqm" event={"ID":"5e2467e3-03a7-43d6-806b-0e4f304d7983","Type":"ContainerDied","Data":"aa89d19790253b752a83e82914e79b50993bb43bfe09cc7f656a09f6c250516d"} Oct 11 04:19:04 crc kubenswrapper[4967]: I1011 04:19:04.580584 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6pmqm" event={"ID":"5e2467e3-03a7-43d6-806b-0e4f304d7983","Type":"ContainerStarted","Data":"3718fd2ad75db9c399f0afd2221358cc9a4ccc59d5ef6e1afac7ebaf670abca0"} Oct 11 04:19:04 crc kubenswrapper[4967]: I1011 04:19:04.602182 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-6pmqm" podStartSLOduration=3.124167978 podStartE2EDuration="5.602166183s" podCreationTimestamp="2025-10-11 04:18:59 +0000 UTC" firstStartedPulling="2025-10-11 04:19:01.549528575 +0000 UTC m=+1669.512737508" lastFinishedPulling="2025-10-11 04:19:04.02752678 +0000 UTC m=+1671.990735713" observedRunningTime="2025-10-11 04:19:04.600799451 +0000 UTC m=+1672.564008404" watchObservedRunningTime="2025-10-11 04:19:04.602166183 +0000 UTC m=+1672.565375116" Oct 11 04:19:06 crc kubenswrapper[4967]: I1011 04:19:06.034589 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-652d-account-create-djxcs"] Oct 11 04:19:06 crc kubenswrapper[4967]: I1011 04:19:06.041060 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-652d-account-create-djxcs"] Oct 11 04:19:06 crc kubenswrapper[4967]: I1011 04:19:06.825335 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="917decdc-0498-4bf6-a45f-cba451199483" path="/var/lib/kubelet/pods/917decdc-0498-4bf6-a45f-cba451199483/volumes" Oct 11 04:19:10 crc kubenswrapper[4967]: I1011 04:19:10.116291 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-6pmqm" Oct 11 04:19:10 crc kubenswrapper[4967]: I1011 04:19:10.117622 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-6pmqm" Oct 11 04:19:10 crc kubenswrapper[4967]: I1011 04:19:10.168791 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-6pmqm" Oct 11 04:19:10 crc kubenswrapper[4967]: I1011 04:19:10.677880 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-6pmqm" Oct 11 04:19:10 crc kubenswrapper[4967]: I1011 04:19:10.723233 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6pmqm"] Oct 11 04:19:12 crc kubenswrapper[4967]: I1011 04:19:12.652121 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-6pmqm" podUID="5e2467e3-03a7-43d6-806b-0e4f304d7983" containerName="registry-server" containerID="cri-o://3718fd2ad75db9c399f0afd2221358cc9a4ccc59d5ef6e1afac7ebaf670abca0" gracePeriod=2 Oct 11 04:19:12 crc kubenswrapper[4967]: I1011 04:19:12.845599 4967 scope.go:117] "RemoveContainer" containerID="6010a3358387442fdc0ff1dfa8bf2d3fb285ef11cfb5f18ad38fe86548ca1f20" Oct 11 04:19:12 crc kubenswrapper[4967]: E1011 04:19:12.860456 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:19:13 crc kubenswrapper[4967]: I1011 04:19:13.112632 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6pmqm" Oct 11 04:19:13 crc kubenswrapper[4967]: I1011 04:19:13.189015 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e2467e3-03a7-43d6-806b-0e4f304d7983-catalog-content\") pod \"5e2467e3-03a7-43d6-806b-0e4f304d7983\" (UID: \"5e2467e3-03a7-43d6-806b-0e4f304d7983\") " Oct 11 04:19:13 crc kubenswrapper[4967]: I1011 04:19:13.189408 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cgjj4\" (UniqueName: \"kubernetes.io/projected/5e2467e3-03a7-43d6-806b-0e4f304d7983-kube-api-access-cgjj4\") pod \"5e2467e3-03a7-43d6-806b-0e4f304d7983\" (UID: \"5e2467e3-03a7-43d6-806b-0e4f304d7983\") " Oct 11 04:19:13 crc kubenswrapper[4967]: I1011 04:19:13.189629 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e2467e3-03a7-43d6-806b-0e4f304d7983-utilities\") pod \"5e2467e3-03a7-43d6-806b-0e4f304d7983\" (UID: \"5e2467e3-03a7-43d6-806b-0e4f304d7983\") " Oct 11 04:19:13 crc kubenswrapper[4967]: I1011 04:19:13.190340 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e2467e3-03a7-43d6-806b-0e4f304d7983-utilities" (OuterVolumeSpecName: "utilities") pod "5e2467e3-03a7-43d6-806b-0e4f304d7983" (UID: "5e2467e3-03a7-43d6-806b-0e4f304d7983"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:19:13 crc kubenswrapper[4967]: I1011 04:19:13.198234 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e2467e3-03a7-43d6-806b-0e4f304d7983-kube-api-access-cgjj4" (OuterVolumeSpecName: "kube-api-access-cgjj4") pod "5e2467e3-03a7-43d6-806b-0e4f304d7983" (UID: "5e2467e3-03a7-43d6-806b-0e4f304d7983"). InnerVolumeSpecName "kube-api-access-cgjj4". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:19:13 crc kubenswrapper[4967]: I1011 04:19:13.237428 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5e2467e3-03a7-43d6-806b-0e4f304d7983-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5e2467e3-03a7-43d6-806b-0e4f304d7983" (UID: "5e2467e3-03a7-43d6-806b-0e4f304d7983"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:19:13 crc kubenswrapper[4967]: I1011 04:19:13.292216 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5e2467e3-03a7-43d6-806b-0e4f304d7983-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 04:19:13 crc kubenswrapper[4967]: I1011 04:19:13.292247 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cgjj4\" (UniqueName: \"kubernetes.io/projected/5e2467e3-03a7-43d6-806b-0e4f304d7983-kube-api-access-cgjj4\") on node \"crc\" DevicePath \"\"" Oct 11 04:19:13 crc kubenswrapper[4967]: I1011 04:19:13.292260 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5e2467e3-03a7-43d6-806b-0e4f304d7983-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 04:19:13 crc kubenswrapper[4967]: I1011 04:19:13.665769 4967 generic.go:334] "Generic (PLEG): container finished" podID="5e2467e3-03a7-43d6-806b-0e4f304d7983" containerID="3718fd2ad75db9c399f0afd2221358cc9a4ccc59d5ef6e1afac7ebaf670abca0" exitCode=0 Oct 11 04:19:13 crc kubenswrapper[4967]: I1011 04:19:13.665816 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6pmqm" event={"ID":"5e2467e3-03a7-43d6-806b-0e4f304d7983","Type":"ContainerDied","Data":"3718fd2ad75db9c399f0afd2221358cc9a4ccc59d5ef6e1afac7ebaf670abca0"} Oct 11 04:19:13 crc kubenswrapper[4967]: I1011 04:19:13.665844 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-6pmqm" event={"ID":"5e2467e3-03a7-43d6-806b-0e4f304d7983","Type":"ContainerDied","Data":"715b02e7e50cf7a922f2674d946f8452d05d0ef556bbfe8a182c9596e532f0ef"} Oct 11 04:19:13 crc kubenswrapper[4967]: I1011 04:19:13.665846 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-6pmqm" Oct 11 04:19:13 crc kubenswrapper[4967]: I1011 04:19:13.665878 4967 scope.go:117] "RemoveContainer" containerID="3718fd2ad75db9c399f0afd2221358cc9a4ccc59d5ef6e1afac7ebaf670abca0" Oct 11 04:19:13 crc kubenswrapper[4967]: I1011 04:19:13.694166 4967 scope.go:117] "RemoveContainer" containerID="aa89d19790253b752a83e82914e79b50993bb43bfe09cc7f656a09f6c250516d" Oct 11 04:19:13 crc kubenswrapper[4967]: I1011 04:19:13.706924 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-6pmqm"] Oct 11 04:19:13 crc kubenswrapper[4967]: I1011 04:19:13.717718 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-6pmqm"] Oct 11 04:19:13 crc kubenswrapper[4967]: I1011 04:19:13.741176 4967 scope.go:117] "RemoveContainer" containerID="3ae5667b452159e6da15dc765036580d6cf5a9bb54f533f07fe0be953390f943" Oct 11 04:19:13 crc kubenswrapper[4967]: I1011 04:19:13.776713 4967 scope.go:117] "RemoveContainer" containerID="3718fd2ad75db9c399f0afd2221358cc9a4ccc59d5ef6e1afac7ebaf670abca0" Oct 11 04:19:13 crc kubenswrapper[4967]: E1011 04:19:13.777160 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3718fd2ad75db9c399f0afd2221358cc9a4ccc59d5ef6e1afac7ebaf670abca0\": container with ID starting with 3718fd2ad75db9c399f0afd2221358cc9a4ccc59d5ef6e1afac7ebaf670abca0 not found: ID does not exist" containerID="3718fd2ad75db9c399f0afd2221358cc9a4ccc59d5ef6e1afac7ebaf670abca0" Oct 11 04:19:13 crc kubenswrapper[4967]: I1011 04:19:13.777192 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3718fd2ad75db9c399f0afd2221358cc9a4ccc59d5ef6e1afac7ebaf670abca0"} err="failed to get container status \"3718fd2ad75db9c399f0afd2221358cc9a4ccc59d5ef6e1afac7ebaf670abca0\": rpc error: code = NotFound desc = could not find container \"3718fd2ad75db9c399f0afd2221358cc9a4ccc59d5ef6e1afac7ebaf670abca0\": container with ID starting with 3718fd2ad75db9c399f0afd2221358cc9a4ccc59d5ef6e1afac7ebaf670abca0 not found: ID does not exist" Oct 11 04:19:13 crc kubenswrapper[4967]: I1011 04:19:13.777215 4967 scope.go:117] "RemoveContainer" containerID="aa89d19790253b752a83e82914e79b50993bb43bfe09cc7f656a09f6c250516d" Oct 11 04:19:13 crc kubenswrapper[4967]: E1011 04:19:13.777877 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"aa89d19790253b752a83e82914e79b50993bb43bfe09cc7f656a09f6c250516d\": container with ID starting with aa89d19790253b752a83e82914e79b50993bb43bfe09cc7f656a09f6c250516d not found: ID does not exist" containerID="aa89d19790253b752a83e82914e79b50993bb43bfe09cc7f656a09f6c250516d" Oct 11 04:19:13 crc kubenswrapper[4967]: I1011 04:19:13.777904 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"aa89d19790253b752a83e82914e79b50993bb43bfe09cc7f656a09f6c250516d"} err="failed to get container status \"aa89d19790253b752a83e82914e79b50993bb43bfe09cc7f656a09f6c250516d\": rpc error: code = NotFound desc = could not find container \"aa89d19790253b752a83e82914e79b50993bb43bfe09cc7f656a09f6c250516d\": container with ID starting with aa89d19790253b752a83e82914e79b50993bb43bfe09cc7f656a09f6c250516d not found: ID does not exist" Oct 11 04:19:13 crc kubenswrapper[4967]: I1011 04:19:13.777920 4967 scope.go:117] "RemoveContainer" containerID="3ae5667b452159e6da15dc765036580d6cf5a9bb54f533f07fe0be953390f943" Oct 11 04:19:13 crc kubenswrapper[4967]: E1011 04:19:13.778394 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3ae5667b452159e6da15dc765036580d6cf5a9bb54f533f07fe0be953390f943\": container with ID starting with 3ae5667b452159e6da15dc765036580d6cf5a9bb54f533f07fe0be953390f943 not found: ID does not exist" containerID="3ae5667b452159e6da15dc765036580d6cf5a9bb54f533f07fe0be953390f943" Oct 11 04:19:13 crc kubenswrapper[4967]: I1011 04:19:13.778417 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3ae5667b452159e6da15dc765036580d6cf5a9bb54f533f07fe0be953390f943"} err="failed to get container status \"3ae5667b452159e6da15dc765036580d6cf5a9bb54f533f07fe0be953390f943\": rpc error: code = NotFound desc = could not find container \"3ae5667b452159e6da15dc765036580d6cf5a9bb54f533f07fe0be953390f943\": container with ID starting with 3ae5667b452159e6da15dc765036580d6cf5a9bb54f533f07fe0be953390f943 not found: ID does not exist" Oct 11 04:19:14 crc kubenswrapper[4967]: I1011 04:19:14.827281 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5e2467e3-03a7-43d6-806b-0e4f304d7983" path="/var/lib/kubelet/pods/5e2467e3-03a7-43d6-806b-0e4f304d7983/volumes" Oct 11 04:19:16 crc kubenswrapper[4967]: I1011 04:19:16.698111 4967 generic.go:334] "Generic (PLEG): container finished" podID="b9d03cd0-766f-4fde-a0bc-95693e4ae111" containerID="7a838cf131945ab75399cf021780705d1d3db35b4babc83434cd6d5b26cdd25a" exitCode=0 Oct 11 04:19:16 crc kubenswrapper[4967]: I1011 04:19:16.698165 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vt6vn/crc-debug-n46l6" event={"ID":"b9d03cd0-766f-4fde-a0bc-95693e4ae111","Type":"ContainerDied","Data":"7a838cf131945ab75399cf021780705d1d3db35b4babc83434cd6d5b26cdd25a"} Oct 11 04:19:17 crc kubenswrapper[4967]: I1011 04:19:17.045639 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-api-abbb-account-create-szgg7"] Oct 11 04:19:17 crc kubenswrapper[4967]: I1011 04:19:17.059382 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-api-abbb-account-create-szgg7"] Oct 11 04:19:17 crc kubenswrapper[4967]: I1011 04:19:17.834754 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vt6vn/crc-debug-n46l6" Oct 11 04:19:17 crc kubenswrapper[4967]: I1011 04:19:17.873526 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-vt6vn/crc-debug-n46l6"] Oct 11 04:19:17 crc kubenswrapper[4967]: I1011 04:19:17.879618 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9d03cd0-766f-4fde-a0bc-95693e4ae111-host\") pod \"b9d03cd0-766f-4fde-a0bc-95693e4ae111\" (UID: \"b9d03cd0-766f-4fde-a0bc-95693e4ae111\") " Oct 11 04:19:17 crc kubenswrapper[4967]: I1011 04:19:17.879688 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h5hld\" (UniqueName: \"kubernetes.io/projected/b9d03cd0-766f-4fde-a0bc-95693e4ae111-kube-api-access-h5hld\") pod \"b9d03cd0-766f-4fde-a0bc-95693e4ae111\" (UID: \"b9d03cd0-766f-4fde-a0bc-95693e4ae111\") " Oct 11 04:19:17 crc kubenswrapper[4967]: I1011 04:19:17.879743 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/b9d03cd0-766f-4fde-a0bc-95693e4ae111-host" (OuterVolumeSpecName: "host") pod "b9d03cd0-766f-4fde-a0bc-95693e4ae111" (UID: "b9d03cd0-766f-4fde-a0bc-95693e4ae111"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 04:19:17 crc kubenswrapper[4967]: I1011 04:19:17.880144 4967 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/b9d03cd0-766f-4fde-a0bc-95693e4ae111-host\") on node \"crc\" DevicePath \"\"" Oct 11 04:19:17 crc kubenswrapper[4967]: I1011 04:19:17.880471 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-vt6vn/crc-debug-n46l6"] Oct 11 04:19:17 crc kubenswrapper[4967]: I1011 04:19:17.887408 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b9d03cd0-766f-4fde-a0bc-95693e4ae111-kube-api-access-h5hld" (OuterVolumeSpecName: "kube-api-access-h5hld") pod "b9d03cd0-766f-4fde-a0bc-95693e4ae111" (UID: "b9d03cd0-766f-4fde-a0bc-95693e4ae111"). InnerVolumeSpecName "kube-api-access-h5hld". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:19:17 crc kubenswrapper[4967]: I1011 04:19:17.981607 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h5hld\" (UniqueName: \"kubernetes.io/projected/b9d03cd0-766f-4fde-a0bc-95693e4ae111-kube-api-access-h5hld\") on node \"crc\" DevicePath \"\"" Oct 11 04:19:18 crc kubenswrapper[4967]: I1011 04:19:18.721913 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c9c3c8f2053aed4914b39a53642db1ff6539bab6315e40a5382f8b5c966c702f" Oct 11 04:19:18 crc kubenswrapper[4967]: I1011 04:19:18.721985 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vt6vn/crc-debug-n46l6" Oct 11 04:19:18 crc kubenswrapper[4967]: I1011 04:19:18.832196 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b9d03cd0-766f-4fde-a0bc-95693e4ae111" path="/var/lib/kubelet/pods/b9d03cd0-766f-4fde-a0bc-95693e4ae111/volumes" Oct 11 04:19:18 crc kubenswrapper[4967]: I1011 04:19:18.832869 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cb662b34-1e00-4cbb-a989-347e80fe9886" path="/var/lib/kubelet/pods/cb662b34-1e00-4cbb-a989-347e80fe9886/volumes" Oct 11 04:19:19 crc kubenswrapper[4967]: I1011 04:19:19.093773 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-vt6vn/crc-debug-mn2x6"] Oct 11 04:19:19 crc kubenswrapper[4967]: E1011 04:19:19.094363 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e2467e3-03a7-43d6-806b-0e4f304d7983" containerName="extract-content" Oct 11 04:19:19 crc kubenswrapper[4967]: I1011 04:19:19.094381 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e2467e3-03a7-43d6-806b-0e4f304d7983" containerName="extract-content" Oct 11 04:19:19 crc kubenswrapper[4967]: E1011 04:19:19.094401 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="b9d03cd0-766f-4fde-a0bc-95693e4ae111" containerName="container-00" Oct 11 04:19:19 crc kubenswrapper[4967]: I1011 04:19:19.094409 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="b9d03cd0-766f-4fde-a0bc-95693e4ae111" containerName="container-00" Oct 11 04:19:19 crc kubenswrapper[4967]: E1011 04:19:19.094423 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e2467e3-03a7-43d6-806b-0e4f304d7983" containerName="extract-utilities" Oct 11 04:19:19 crc kubenswrapper[4967]: I1011 04:19:19.094432 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e2467e3-03a7-43d6-806b-0e4f304d7983" containerName="extract-utilities" Oct 11 04:19:19 crc kubenswrapper[4967]: E1011 04:19:19.094451 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e2467e3-03a7-43d6-806b-0e4f304d7983" containerName="registry-server" Oct 11 04:19:19 crc kubenswrapper[4967]: I1011 04:19:19.094459 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e2467e3-03a7-43d6-806b-0e4f304d7983" containerName="registry-server" Oct 11 04:19:19 crc kubenswrapper[4967]: I1011 04:19:19.094709 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e2467e3-03a7-43d6-806b-0e4f304d7983" containerName="registry-server" Oct 11 04:19:19 crc kubenswrapper[4967]: I1011 04:19:19.094729 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="b9d03cd0-766f-4fde-a0bc-95693e4ae111" containerName="container-00" Oct 11 04:19:19 crc kubenswrapper[4967]: I1011 04:19:19.095565 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vt6vn/crc-debug-mn2x6" Oct 11 04:19:19 crc kubenswrapper[4967]: I1011 04:19:19.207961 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h9msz\" (UniqueName: \"kubernetes.io/projected/2a074bda-bf71-4c40-97a7-4b180956e0b1-kube-api-access-h9msz\") pod \"crc-debug-mn2x6\" (UID: \"2a074bda-bf71-4c40-97a7-4b180956e0b1\") " pod="openshift-must-gather-vt6vn/crc-debug-mn2x6" Oct 11 04:19:19 crc kubenswrapper[4967]: I1011 04:19:19.208319 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2a074bda-bf71-4c40-97a7-4b180956e0b1-host\") pod \"crc-debug-mn2x6\" (UID: \"2a074bda-bf71-4c40-97a7-4b180956e0b1\") " pod="openshift-must-gather-vt6vn/crc-debug-mn2x6" Oct 11 04:19:19 crc kubenswrapper[4967]: I1011 04:19:19.311343 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h9msz\" (UniqueName: \"kubernetes.io/projected/2a074bda-bf71-4c40-97a7-4b180956e0b1-kube-api-access-h9msz\") pod \"crc-debug-mn2x6\" (UID: \"2a074bda-bf71-4c40-97a7-4b180956e0b1\") " pod="openshift-must-gather-vt6vn/crc-debug-mn2x6" Oct 11 04:19:19 crc kubenswrapper[4967]: I1011 04:19:19.311574 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2a074bda-bf71-4c40-97a7-4b180956e0b1-host\") pod \"crc-debug-mn2x6\" (UID: \"2a074bda-bf71-4c40-97a7-4b180956e0b1\") " pod="openshift-must-gather-vt6vn/crc-debug-mn2x6" Oct 11 04:19:19 crc kubenswrapper[4967]: I1011 04:19:19.311717 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2a074bda-bf71-4c40-97a7-4b180956e0b1-host\") pod \"crc-debug-mn2x6\" (UID: \"2a074bda-bf71-4c40-97a7-4b180956e0b1\") " pod="openshift-must-gather-vt6vn/crc-debug-mn2x6" Oct 11 04:19:19 crc kubenswrapper[4967]: I1011 04:19:19.348454 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h9msz\" (UniqueName: \"kubernetes.io/projected/2a074bda-bf71-4c40-97a7-4b180956e0b1-kube-api-access-h9msz\") pod \"crc-debug-mn2x6\" (UID: \"2a074bda-bf71-4c40-97a7-4b180956e0b1\") " pod="openshift-must-gather-vt6vn/crc-debug-mn2x6" Oct 11 04:19:19 crc kubenswrapper[4967]: I1011 04:19:19.413061 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vt6vn/crc-debug-mn2x6" Oct 11 04:19:19 crc kubenswrapper[4967]: I1011 04:19:19.738296 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vt6vn/crc-debug-mn2x6" event={"ID":"2a074bda-bf71-4c40-97a7-4b180956e0b1","Type":"ContainerStarted","Data":"aaf10915876775ab06c02796a259fdd767edb2dc8371c87aa795b922be5a13be"} Oct 11 04:19:19 crc kubenswrapper[4967]: I1011 04:19:19.738578 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vt6vn/crc-debug-mn2x6" event={"ID":"2a074bda-bf71-4c40-97a7-4b180956e0b1","Type":"ContainerStarted","Data":"396148bd17efffd3b8cccc08f621181ea8d0e60f64bddfd986330eeab06413f1"} Oct 11 04:19:19 crc kubenswrapper[4967]: I1011 04:19:19.757136 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-vt6vn/crc-debug-mn2x6" podStartSLOduration=0.757117523 podStartE2EDuration="757.117523ms" podCreationTimestamp="2025-10-11 04:19:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:19:19.756634141 +0000 UTC m=+1687.719843124" watchObservedRunningTime="2025-10-11 04:19:19.757117523 +0000 UTC m=+1687.720326446" Oct 11 04:19:20 crc kubenswrapper[4967]: I1011 04:19:20.748295 4967 generic.go:334] "Generic (PLEG): container finished" podID="2a074bda-bf71-4c40-97a7-4b180956e0b1" containerID="aaf10915876775ab06c02796a259fdd767edb2dc8371c87aa795b922be5a13be" exitCode=0 Oct 11 04:19:20 crc kubenswrapper[4967]: I1011 04:19:20.748369 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vt6vn/crc-debug-mn2x6" event={"ID":"2a074bda-bf71-4c40-97a7-4b180956e0b1","Type":"ContainerDied","Data":"aaf10915876775ab06c02796a259fdd767edb2dc8371c87aa795b922be5a13be"} Oct 11 04:19:21 crc kubenswrapper[4967]: I1011 04:19:21.267597 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-84674df998-htlks_c72a2b43-59ac-450c-87d3-f16c9ee4afa0/barbican-api-log/0.log" Oct 11 04:19:21 crc kubenswrapper[4967]: I1011 04:19:21.269590 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-84674df998-htlks_c72a2b43-59ac-450c-87d3-f16c9ee4afa0/barbican-api/0.log" Oct 11 04:19:21 crc kubenswrapper[4967]: I1011 04:19:21.458333 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7769b9cccd-zkh5v_38e5cf89-c793-4207-b06e-d117da3dd152/barbican-keystone-listener/0.log" Oct 11 04:19:21 crc kubenswrapper[4967]: I1011 04:19:21.614993 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7769b9cccd-zkh5v_38e5cf89-c793-4207-b06e-d117da3dd152/barbican-keystone-listener-log/0.log" Oct 11 04:19:21 crc kubenswrapper[4967]: I1011 04:19:21.678940 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6c87f65cff-p6j2x_fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35/barbican-worker/0.log" Oct 11 04:19:21 crc kubenswrapper[4967]: I1011 04:19:21.872730 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vt6vn/crc-debug-mn2x6" Oct 11 04:19:21 crc kubenswrapper[4967]: I1011 04:19:21.881402 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6c87f65cff-p6j2x_fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35/barbican-worker-log/0.log" Oct 11 04:19:21 crc kubenswrapper[4967]: I1011 04:19:21.907590 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-vt6vn/crc-debug-mn2x6"] Oct 11 04:19:21 crc kubenswrapper[4967]: I1011 04:19:21.917379 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-vt6vn/crc-debug-mn2x6"] Oct 11 04:19:21 crc kubenswrapper[4967]: I1011 04:19:21.940409 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c47c6348-b68f-410e-93d7-c10eed165a34/ceilometer-central-agent/0.log" Oct 11 04:19:21 crc kubenswrapper[4967]: I1011 04:19:21.956480 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h9msz\" (UniqueName: \"kubernetes.io/projected/2a074bda-bf71-4c40-97a7-4b180956e0b1-kube-api-access-h9msz\") pod \"2a074bda-bf71-4c40-97a7-4b180956e0b1\" (UID: \"2a074bda-bf71-4c40-97a7-4b180956e0b1\") " Oct 11 04:19:21 crc kubenswrapper[4967]: I1011 04:19:21.956576 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2a074bda-bf71-4c40-97a7-4b180956e0b1-host\") pod \"2a074bda-bf71-4c40-97a7-4b180956e0b1\" (UID: \"2a074bda-bf71-4c40-97a7-4b180956e0b1\") " Oct 11 04:19:21 crc kubenswrapper[4967]: I1011 04:19:21.956717 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/2a074bda-bf71-4c40-97a7-4b180956e0b1-host" (OuterVolumeSpecName: "host") pod "2a074bda-bf71-4c40-97a7-4b180956e0b1" (UID: "2a074bda-bf71-4c40-97a7-4b180956e0b1"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 04:19:21 crc kubenswrapper[4967]: I1011 04:19:21.957155 4967 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/2a074bda-bf71-4c40-97a7-4b180956e0b1-host\") on node \"crc\" DevicePath \"\"" Oct 11 04:19:21 crc kubenswrapper[4967]: I1011 04:19:21.982212 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2a074bda-bf71-4c40-97a7-4b180956e0b1-kube-api-access-h9msz" (OuterVolumeSpecName: "kube-api-access-h9msz") pod "2a074bda-bf71-4c40-97a7-4b180956e0b1" (UID: "2a074bda-bf71-4c40-97a7-4b180956e0b1"). InnerVolumeSpecName "kube-api-access-h9msz". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:19:22 crc kubenswrapper[4967]: I1011 04:19:22.059065 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h9msz\" (UniqueName: \"kubernetes.io/projected/2a074bda-bf71-4c40-97a7-4b180956e0b1-kube-api-access-h9msz\") on node \"crc\" DevicePath \"\"" Oct 11 04:19:22 crc kubenswrapper[4967]: I1011 04:19:22.074605 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c47c6348-b68f-410e-93d7-c10eed165a34/ceilometer-notification-agent/0.log" Oct 11 04:19:22 crc kubenswrapper[4967]: I1011 04:19:22.131211 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c47c6348-b68f-410e-93d7-c10eed165a34/sg-core/0.log" Oct 11 04:19:22 crc kubenswrapper[4967]: I1011 04:19:22.165855 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c47c6348-b68f-410e-93d7-c10eed165a34/proxy-httpd/0.log" Oct 11 04:19:22 crc kubenswrapper[4967]: I1011 04:19:22.305538 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_c900849f-54f8-44ed-b759-ae6fbbe441da/cinder-api/0.log" Oct 11 04:19:22 crc kubenswrapper[4967]: I1011 04:19:22.387291 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_c900849f-54f8-44ed-b759-ae6fbbe441da/cinder-api-log/0.log" Oct 11 04:19:22 crc kubenswrapper[4967]: I1011 04:19:22.526113 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_a824a183-d65e-4b45-8ecf-72bd3ce9063b/cinder-scheduler/0.log" Oct 11 04:19:22 crc kubenswrapper[4967]: I1011 04:19:22.603639 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_a824a183-d65e-4b45-8ecf-72bd3ce9063b/probe/0.log" Oct 11 04:19:22 crc kubenswrapper[4967]: I1011 04:19:22.733130 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cd5cbd7b9-87xhd_ef02fdf0-4696-42b9-9095-b000a70fe334/init/0.log" Oct 11 04:19:22 crc kubenswrapper[4967]: I1011 04:19:22.763962 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="396148bd17efffd3b8cccc08f621181ea8d0e60f64bddfd986330eeab06413f1" Oct 11 04:19:22 crc kubenswrapper[4967]: I1011 04:19:22.764024 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vt6vn/crc-debug-mn2x6" Oct 11 04:19:22 crc kubenswrapper[4967]: I1011 04:19:22.824744 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2a074bda-bf71-4c40-97a7-4b180956e0b1" path="/var/lib/kubelet/pods/2a074bda-bf71-4c40-97a7-4b180956e0b1/volumes" Oct 11 04:19:22 crc kubenswrapper[4967]: I1011 04:19:22.899988 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cd5cbd7b9-87xhd_ef02fdf0-4696-42b9-9095-b000a70fe334/init/0.log" Oct 11 04:19:22 crc kubenswrapper[4967]: I1011 04:19:22.938375 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cd5cbd7b9-87xhd_ef02fdf0-4696-42b9-9095-b000a70fe334/dnsmasq-dns/0.log" Oct 11 04:19:23 crc kubenswrapper[4967]: I1011 04:19:23.024675 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-1117-account-create-76qc9"] Oct 11 04:19:23 crc kubenswrapper[4967]: I1011 04:19:23.035404 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-1117-account-create-76qc9"] Oct 11 04:19:23 crc kubenswrapper[4967]: I1011 04:19:23.092677 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_662b8e47-db97-4003-892a-48c25cba4589/glance-httpd/0.log" Oct 11 04:19:23 crc kubenswrapper[4967]: I1011 04:19:23.140886 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-vt6vn/crc-debug-6mwll"] Oct 11 04:19:23 crc kubenswrapper[4967]: E1011 04:19:23.141240 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2a074bda-bf71-4c40-97a7-4b180956e0b1" containerName="container-00" Oct 11 04:19:23 crc kubenswrapper[4967]: I1011 04:19:23.141258 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="2a074bda-bf71-4c40-97a7-4b180956e0b1" containerName="container-00" Oct 11 04:19:23 crc kubenswrapper[4967]: I1011 04:19:23.141462 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="2a074bda-bf71-4c40-97a7-4b180956e0b1" containerName="container-00" Oct 11 04:19:23 crc kubenswrapper[4967]: I1011 04:19:23.142038 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vt6vn/crc-debug-6mwll" Oct 11 04:19:23 crc kubenswrapper[4967]: I1011 04:19:23.160456 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_662b8e47-db97-4003-892a-48c25cba4589/glance-log/0.log" Oct 11 04:19:23 crc kubenswrapper[4967]: I1011 04:19:23.179624 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-spc4g\" (UniqueName: \"kubernetes.io/projected/54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0-kube-api-access-spc4g\") pod \"crc-debug-6mwll\" (UID: \"54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0\") " pod="openshift-must-gather-vt6vn/crc-debug-6mwll" Oct 11 04:19:23 crc kubenswrapper[4967]: I1011 04:19:23.179779 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0-host\") pod \"crc-debug-6mwll\" (UID: \"54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0\") " pod="openshift-must-gather-vt6vn/crc-debug-6mwll" Oct 11 04:19:23 crc kubenswrapper[4967]: I1011 04:19:23.281434 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-spc4g\" (UniqueName: \"kubernetes.io/projected/54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0-kube-api-access-spc4g\") pod \"crc-debug-6mwll\" (UID: \"54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0\") " pod="openshift-must-gather-vt6vn/crc-debug-6mwll" Oct 11 04:19:23 crc kubenswrapper[4967]: I1011 04:19:23.281554 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0-host\") pod \"crc-debug-6mwll\" (UID: \"54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0\") " pod="openshift-must-gather-vt6vn/crc-debug-6mwll" Oct 11 04:19:23 crc kubenswrapper[4967]: I1011 04:19:23.281681 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0-host\") pod \"crc-debug-6mwll\" (UID: \"54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0\") " pod="openshift-must-gather-vt6vn/crc-debug-6mwll" Oct 11 04:19:23 crc kubenswrapper[4967]: I1011 04:19:23.282457 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_53708cdd-1297-4302-9d8a-e00ba57c01b3/glance-httpd/0.log" Oct 11 04:19:23 crc kubenswrapper[4967]: I1011 04:19:23.301241 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-spc4g\" (UniqueName: \"kubernetes.io/projected/54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0-kube-api-access-spc4g\") pod \"crc-debug-6mwll\" (UID: \"54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0\") " pod="openshift-must-gather-vt6vn/crc-debug-6mwll" Oct 11 04:19:23 crc kubenswrapper[4967]: I1011 04:19:23.372343 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_53708cdd-1297-4302-9d8a-e00ba57c01b3/glance-log/0.log" Oct 11 04:19:23 crc kubenswrapper[4967]: I1011 04:19:23.461801 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vt6vn/crc-debug-6mwll" Oct 11 04:19:23 crc kubenswrapper[4967]: I1011 04:19:23.520402 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-85b58c7f48-9dgpg_f8ec340a-d11c-4b3b-9d2c-691df43b8892/init/0.log" Oct 11 04:19:23 crc kubenswrapper[4967]: I1011 04:19:23.684089 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-85b58c7f48-9dgpg_f8ec340a-d11c-4b3b-9d2c-691df43b8892/init/0.log" Oct 11 04:19:23 crc kubenswrapper[4967]: I1011 04:19:23.739406 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-85b58c7f48-9dgpg_f8ec340a-d11c-4b3b-9d2c-691df43b8892/ironic-api-log/0.log" Oct 11 04:19:23 crc kubenswrapper[4967]: I1011 04:19:23.779717 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vt6vn/crc-debug-6mwll" event={"ID":"54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0","Type":"ContainerStarted","Data":"d33d963b2aacb7cc3ccea107ae01569eef1f0d5ed02980c627b0f04b14c961a4"} Oct 11 04:19:23 crc kubenswrapper[4967]: I1011 04:19:23.779774 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vt6vn/crc-debug-6mwll" event={"ID":"54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0","Type":"ContainerStarted","Data":"889ab292d6e5f5d2500dfeaa37fa3c68bbd324348ef206cc4ec44dd6ca373b0b"} Oct 11 04:19:23 crc kubenswrapper[4967]: I1011 04:19:23.786431 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-85b58c7f48-9dgpg_f8ec340a-d11c-4b3b-9d2c-691df43b8892/ironic-api/0.log" Oct 11 04:19:23 crc kubenswrapper[4967]: I1011 04:19:23.819770 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-vt6vn/crc-debug-6mwll"] Oct 11 04:19:23 crc kubenswrapper[4967]: I1011 04:19:23.830261 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-vt6vn/crc-debug-6mwll"] Oct 11 04:19:23 crc kubenswrapper[4967]: I1011 04:19:23.932344 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_64c75ea6-aed6-4bf6-8eef-39e76384e14f/init/0.log" Oct 11 04:19:24 crc kubenswrapper[4967]: I1011 04:19:24.149670 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_64c75ea6-aed6-4bf6-8eef-39e76384e14f/ironic-python-agent-init/0.log" Oct 11 04:19:24 crc kubenswrapper[4967]: I1011 04:19:24.185057 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_64c75ea6-aed6-4bf6-8eef-39e76384e14f/init/0.log" Oct 11 04:19:24 crc kubenswrapper[4967]: I1011 04:19:24.223316 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_64c75ea6-aed6-4bf6-8eef-39e76384e14f/ironic-python-agent-init/0.log" Oct 11 04:19:24 crc kubenswrapper[4967]: I1011 04:19:24.448484 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_64c75ea6-aed6-4bf6-8eef-39e76384e14f/init/0.log" Oct 11 04:19:24 crc kubenswrapper[4967]: I1011 04:19:24.516025 4967 scope.go:117] "RemoveContainer" containerID="f8b2adf251e6a745e947a0a7231014ac44446c459e8fa03fa0dc9d372d3475dc" Oct 11 04:19:24 crc kubenswrapper[4967]: I1011 04:19:24.565570 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_64c75ea6-aed6-4bf6-8eef-39e76384e14f/ironic-python-agent-init/0.log" Oct 11 04:19:24 crc kubenswrapper[4967]: I1011 04:19:24.801524 4967 generic.go:334] "Generic (PLEG): container finished" podID="54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0" containerID="d33d963b2aacb7cc3ccea107ae01569eef1f0d5ed02980c627b0f04b14c961a4" exitCode=0 Oct 11 04:19:24 crc kubenswrapper[4967]: I1011 04:19:24.820537 4967 scope.go:117] "RemoveContainer" containerID="6010a3358387442fdc0ff1dfa8bf2d3fb285ef11cfb5f18ad38fe86548ca1f20" Oct 11 04:19:24 crc kubenswrapper[4967]: E1011 04:19:24.820932 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:19:24 crc kubenswrapper[4967]: I1011 04:19:24.821721 4967 scope.go:117] "RemoveContainer" containerID="383dd9d4109b281e5b663d9eeb8e34c15b3a8486d8b06d354d56d1399811c062" Oct 11 04:19:24 crc kubenswrapper[4967]: I1011 04:19:24.834400 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a758d270-0d47-415f-bae1-ae7288e90245" path="/var/lib/kubelet/pods/a758d270-0d47-415f-bae1-ae7288e90245/volumes" Oct 11 04:19:24 crc kubenswrapper[4967]: I1011 04:19:24.926153 4967 scope.go:117] "RemoveContainer" containerID="5c3a8c7a53f8a0661cc2d27adc0254e52ee01bdf4383ead7500f920e6d5de817" Oct 11 04:19:24 crc kubenswrapper[4967]: I1011 04:19:24.959394 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_64c75ea6-aed6-4bf6-8eef-39e76384e14f/init/0.log" Oct 11 04:19:24 crc kubenswrapper[4967]: I1011 04:19:24.960093 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vt6vn/crc-debug-6mwll" Oct 11 04:19:24 crc kubenswrapper[4967]: I1011 04:19:24.970309 4967 scope.go:117] "RemoveContainer" containerID="e018e7db55ec5edeecce65b93021dd2c61dd64dd6b73dd0f4ea82c4c455246a7" Oct 11 04:19:25 crc kubenswrapper[4967]: I1011 04:19:25.000512 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_64c75ea6-aed6-4bf6-8eef-39e76384e14f/ironic-python-agent-init/0.log" Oct 11 04:19:25 crc kubenswrapper[4967]: I1011 04:19:25.011176 4967 scope.go:117] "RemoveContainer" containerID="82f3a6bd6722ea37f904a5fbd419d21d038cfd670bba6e8fa7206aed09197121" Oct 11 04:19:25 crc kubenswrapper[4967]: I1011 04:19:25.025728 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0-host\") pod \"54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0\" (UID: \"54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0\") " Oct 11 04:19:25 crc kubenswrapper[4967]: I1011 04:19:25.025925 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-spc4g\" (UniqueName: \"kubernetes.io/projected/54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0-kube-api-access-spc4g\") pod \"54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0\" (UID: \"54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0\") " Oct 11 04:19:25 crc kubenswrapper[4967]: I1011 04:19:25.030344 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0-host" (OuterVolumeSpecName: "host") pod "54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0" (UID: "54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 04:19:25 crc kubenswrapper[4967]: I1011 04:19:25.032965 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0-kube-api-access-spc4g" (OuterVolumeSpecName: "kube-api-access-spc4g") pod "54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0" (UID: "54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0"). InnerVolumeSpecName "kube-api-access-spc4g". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:19:25 crc kubenswrapper[4967]: I1011 04:19:25.044747 4967 scope.go:117] "RemoveContainer" containerID="de98430f092cd5ed87090b453397b430785e49854d3b7844ef1aa6d36fcf2da2" Oct 11 04:19:25 crc kubenswrapper[4967]: I1011 04:19:25.052419 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_64c75ea6-aed6-4bf6-8eef-39e76384e14f/pxe-init/0.log" Oct 11 04:19:25 crc kubenswrapper[4967]: I1011 04:19:25.072670 4967 scope.go:117] "RemoveContainer" containerID="dc95140e2858d0bdbe1c8a9b46e69c6ce135d99beee50963bbfde3504a87caf6" Oct 11 04:19:25 crc kubenswrapper[4967]: I1011 04:19:25.111478 4967 scope.go:117] "RemoveContainer" containerID="433cef11bc0a2150c9ca5cf11665083f8de67f71d9f8a86c58ef07e0b0409f7e" Oct 11 04:19:25 crc kubenswrapper[4967]: I1011 04:19:25.129366 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-spc4g\" (UniqueName: \"kubernetes.io/projected/54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0-kube-api-access-spc4g\") on node \"crc\" DevicePath \"\"" Oct 11 04:19:25 crc kubenswrapper[4967]: I1011 04:19:25.129396 4967 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0-host\") on node \"crc\" DevicePath \"\"" Oct 11 04:19:25 crc kubenswrapper[4967]: I1011 04:19:25.153300 4967 scope.go:117] "RemoveContainer" containerID="478a3e8f1b62c6faf971349d240c3f760ef34f0e4761a46e75a50421e9b3abe6" Oct 11 04:19:25 crc kubenswrapper[4967]: I1011 04:19:25.392370 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_64c75ea6-aed6-4bf6-8eef-39e76384e14f/httpboot/0.log" Oct 11 04:19:25 crc kubenswrapper[4967]: I1011 04:19:25.480897 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_64c75ea6-aed6-4bf6-8eef-39e76384e14f/ironic-conductor/0.log" Oct 11 04:19:25 crc kubenswrapper[4967]: I1011 04:19:25.555654 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_64c75ea6-aed6-4bf6-8eef-39e76384e14f/ramdisk-logs/0.log" Oct 11 04:19:25 crc kubenswrapper[4967]: I1011 04:19:25.745595 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-db-sync-sdg2s_1d249b36-580e-424d-8f36-e366c932411c/init/0.log" Oct 11 04:19:25 crc kubenswrapper[4967]: I1011 04:19:25.820492 4967 scope.go:117] "RemoveContainer" containerID="d33d963b2aacb7cc3ccea107ae01569eef1f0d5ed02980c627b0f04b14c961a4" Oct 11 04:19:25 crc kubenswrapper[4967]: I1011 04:19:25.820850 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vt6vn/crc-debug-6mwll" Oct 11 04:19:25 crc kubenswrapper[4967]: I1011 04:19:25.984216 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-db-sync-sdg2s_1d249b36-580e-424d-8f36-e366c932411c/init/0.log" Oct 11 04:19:25 crc kubenswrapper[4967]: I1011 04:19:25.987523 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-db-sync-sdg2s_1d249b36-580e-424d-8f36-e366c932411c/ironic-db-sync/0.log" Oct 11 04:19:26 crc kubenswrapper[4967]: I1011 04:19:26.070748 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_64c75ea6-aed6-4bf6-8eef-39e76384e14f/pxe-init/0.log" Oct 11 04:19:26 crc kubenswrapper[4967]: I1011 04:19:26.074366 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_64c75ea6-aed6-4bf6-8eef-39e76384e14f/pxe-init/0.log" Oct 11 04:19:26 crc kubenswrapper[4967]: I1011 04:19:26.141830 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_c619427d-62f7-463c-8262-d3b9fdd6ade0/ironic-python-agent-init/0.log" Oct 11 04:19:26 crc kubenswrapper[4967]: I1011 04:19:26.404463 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_c619427d-62f7-463c-8262-d3b9fdd6ade0/ironic-python-agent-init/0.log" Oct 11 04:19:26 crc kubenswrapper[4967]: I1011 04:19:26.405940 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_64c75ea6-aed6-4bf6-8eef-39e76384e14f/pxe-init/0.log" Oct 11 04:19:26 crc kubenswrapper[4967]: I1011 04:19:26.449320 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_c619427d-62f7-463c-8262-d3b9fdd6ade0/inspector-pxe-init/0.log" Oct 11 04:19:26 crc kubenswrapper[4967]: I1011 04:19:26.463671 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_c619427d-62f7-463c-8262-d3b9fdd6ade0/inspector-pxe-init/0.log" Oct 11 04:19:26 crc kubenswrapper[4967]: I1011 04:19:26.550505 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_c619427d-62f7-463c-8262-d3b9fdd6ade0/ironic-python-agent-init/0.log" Oct 11 04:19:26 crc kubenswrapper[4967]: I1011 04:19:26.587657 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_c619427d-62f7-463c-8262-d3b9fdd6ade0/inspector-httpboot/0.log" Oct 11 04:19:26 crc kubenswrapper[4967]: I1011 04:19:26.629940 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_c619427d-62f7-463c-8262-d3b9fdd6ade0/ironic-inspector/0.log" Oct 11 04:19:26 crc kubenswrapper[4967]: I1011 04:19:26.633938 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_c619427d-62f7-463c-8262-d3b9fdd6ade0/inspector-pxe-init/0.log" Oct 11 04:19:26 crc kubenswrapper[4967]: I1011 04:19:26.697115 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_c619427d-62f7-463c-8262-d3b9fdd6ade0/ironic-inspector-httpd/0.log" Oct 11 04:19:26 crc kubenswrapper[4967]: I1011 04:19:26.725983 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_c619427d-62f7-463c-8262-d3b9fdd6ade0/ramdisk-logs/0.log" Oct 11 04:19:26 crc kubenswrapper[4967]: I1011 04:19:26.826084 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0" path="/var/lib/kubelet/pods/54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0/volumes" Oct 11 04:19:26 crc kubenswrapper[4967]: I1011 04:19:26.853893 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-db-sync-m5k2b_1d9749c8-f6a4-497f-b6dd-794865243379/ironic-inspector-db-sync/0.log" Oct 11 04:19:26 crc kubenswrapper[4967]: I1011 04:19:26.876569 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-neutron-agent-657f6548d9-9fb8w_4390a950-7185-4fc0-b152-b5927940032b/ironic-neutron-agent/3.log" Oct 11 04:19:26 crc kubenswrapper[4967]: I1011 04:19:26.943542 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-neutron-agent-657f6548d9-9fb8w_4390a950-7185-4fc0-b152-b5927940032b/ironic-neutron-agent/2.log" Oct 11 04:19:27 crc kubenswrapper[4967]: I1011 04:19:27.145185 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_f5917a28-16f9-4b7b-8b12-04c44c5333cf/kube-state-metrics/0.log" Oct 11 04:19:27 crc kubenswrapper[4967]: I1011 04:19:27.186883 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-76b6798d84-nx4fq_eb7b8d58-ee2e-4203-a1a9-44b2af5599cb/keystone-api/0.log" Oct 11 04:19:27 crc kubenswrapper[4967]: I1011 04:19:27.390039 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5f96fd6969-6srgk_c21d8a72-9113-4b40-9cd2-fa7f56f749cd/neutron-httpd/0.log" Oct 11 04:19:27 crc kubenswrapper[4967]: I1011 04:19:27.489570 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5f96fd6969-6srgk_c21d8a72-9113-4b40-9cd2-fa7f56f749cd/neutron-api/0.log" Oct 11 04:19:27 crc kubenswrapper[4967]: I1011 04:19:27.646793 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_4c3cc259-3802-4172-8ae6-446cfbcd4ca5/nova-api-api/0.log" Oct 11 04:19:27 crc kubenswrapper[4967]: I1011 04:19:27.668763 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_4c3cc259-3802-4172-8ae6-446cfbcd4ca5/nova-api-log/0.log" Oct 11 04:19:27 crc kubenswrapper[4967]: I1011 04:19:27.698539 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-cell-mapping-krtvp_9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e/nova-manage/0.log" Oct 11 04:19:27 crc kubenswrapper[4967]: I1011 04:19:27.911047 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-db-sync-xh6jw_d4a234ae-b210-4ac3-9bf2-841f586ee4a3/nova-cell0-conductor-db-sync/0.log" Oct 11 04:19:27 crc kubenswrapper[4967]: I1011 04:19:27.937617 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_67b30c76-971c-4f10-a692-4bbf2a1e8531/nova-cell0-conductor-conductor/0.log" Oct 11 04:19:28 crc kubenswrapper[4967]: I1011 04:19:28.334913 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-cell-mapping-2nl8n_11ef0366-73db-43a2-88c4-780d2a0084c9/nova-manage/0.log" Oct 11 04:19:28 crc kubenswrapper[4967]: I1011 04:19:28.545727 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_a9b2e413-f97a-4b67-9071-d1aed0db806f/nova-cell1-conductor-conductor/0.log" Oct 11 04:19:28 crc kubenswrapper[4967]: I1011 04:19:28.576488 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-db-sync-dt5p8_702714d0-d935-49ae-b19f-c0f80acd7758/nova-cell1-conductor-db-sync/0.log" Oct 11 04:19:28 crc kubenswrapper[4967]: I1011 04:19:28.801910 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_2914dd42-ea01-44b5-a8cd-c23fc07e3170/nova-cell1-novncproxy-novncproxy/0.log" Oct 11 04:19:28 crc kubenswrapper[4967]: I1011 04:19:28.913877 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_8e701cd5-1cf9-4ed0-87c6-2023c41e7982/nova-metadata-log/0.log" Oct 11 04:19:29 crc kubenswrapper[4967]: I1011 04:19:29.148666 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_388b5da4-c3cd-43f3-b839-cfd0b6c2a8f8/nova-scheduler-scheduler/0.log" Oct 11 04:19:29 crc kubenswrapper[4967]: I1011 04:19:29.229272 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_8e701cd5-1cf9-4ed0-87c6-2023c41e7982/nova-metadata-metadata/0.log" Oct 11 04:19:29 crc kubenswrapper[4967]: I1011 04:19:29.251535 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_2f52a524-225c-4609-8cb0-0ec800e14aad/mysql-bootstrap/0.log" Oct 11 04:19:29 crc kubenswrapper[4967]: I1011 04:19:29.508322 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_2f52a524-225c-4609-8cb0-0ec800e14aad/mysql-bootstrap/0.log" Oct 11 04:19:29 crc kubenswrapper[4967]: I1011 04:19:29.556690 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_2f52a524-225c-4609-8cb0-0ec800e14aad/galera/0.log" Oct 11 04:19:29 crc kubenswrapper[4967]: I1011 04:19:29.580387 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ddc18aa7-8599-40c5-a019-972c13162f83/mysql-bootstrap/0.log" Oct 11 04:19:29 crc kubenswrapper[4967]: I1011 04:19:29.717572 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ddc18aa7-8599-40c5-a019-972c13162f83/mysql-bootstrap/0.log" Oct 11 04:19:29 crc kubenswrapper[4967]: I1011 04:19:29.791519 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_23c5100e-07d0-4897-98de-f4e91329085e/openstackclient/0.log" Oct 11 04:19:29 crc kubenswrapper[4967]: I1011 04:19:29.804750 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ddc18aa7-8599-40c5-a019-972c13162f83/galera/0.log" Oct 11 04:19:29 crc kubenswrapper[4967]: I1011 04:19:29.952311 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-2n27p_6df16c29-dfa6-4972-9b1c-c1190bb1ca2a/openstack-network-exporter/0.log" Oct 11 04:19:30 crc kubenswrapper[4967]: I1011 04:19:30.074542 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7b5rp_4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20/ovsdb-server-init/0.log" Oct 11 04:19:30 crc kubenswrapper[4967]: I1011 04:19:30.225883 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7b5rp_4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20/ovsdb-server-init/0.log" Oct 11 04:19:30 crc kubenswrapper[4967]: I1011 04:19:30.279093 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7b5rp_4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20/ovsdb-server/0.log" Oct 11 04:19:30 crc kubenswrapper[4967]: I1011 04:19:30.282394 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7b5rp_4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20/ovs-vswitchd/0.log" Oct 11 04:19:30 crc kubenswrapper[4967]: I1011 04:19:30.432410 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-wvrnq_acad8def-db3f-40a3-a95d-28b7503d4aae/ovn-controller/0.log" Oct 11 04:19:30 crc kubenswrapper[4967]: I1011 04:19:30.464684 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_942f72f2-4ef8-4056-b470-cbc9b2741f10/openstack-network-exporter/0.log" Oct 11 04:19:30 crc kubenswrapper[4967]: I1011 04:19:30.599974 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_942f72f2-4ef8-4056-b470-cbc9b2741f10/ovn-northd/0.log" Oct 11 04:19:30 crc kubenswrapper[4967]: I1011 04:19:30.690575 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_90a0e0c5-70a6-4330-986c-a09d244f781b/openstack-network-exporter/0.log" Oct 11 04:19:30 crc kubenswrapper[4967]: I1011 04:19:30.765935 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_90a0e0c5-70a6-4330-986c-a09d244f781b/ovsdbserver-nb/0.log" Oct 11 04:19:30 crc kubenswrapper[4967]: I1011 04:19:30.915453 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca/openstack-network-exporter/0.log" Oct 11 04:19:30 crc kubenswrapper[4967]: I1011 04:19:30.974523 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca/ovsdbserver-sb/0.log" Oct 11 04:19:31 crc kubenswrapper[4967]: I1011 04:19:31.120752 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5cdcf9b996-625cn_878b0a76-9c37-4210-870b-5fed33721831/placement-api/0.log" Oct 11 04:19:31 crc kubenswrapper[4967]: I1011 04:19:31.208436 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5cdcf9b996-625cn_878b0a76-9c37-4210-870b-5fed33721831/placement-log/0.log" Oct 11 04:19:31 crc kubenswrapper[4967]: I1011 04:19:31.276305 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_c41a3ead-949a-4f76-9a58-0e1b4db0abb5/setup-container/0.log" Oct 11 04:19:31 crc kubenswrapper[4967]: I1011 04:19:31.631319 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_c41a3ead-949a-4f76-9a58-0e1b4db0abb5/setup-container/0.log" Oct 11 04:19:31 crc kubenswrapper[4967]: I1011 04:19:31.687477 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab/setup-container/0.log" Oct 11 04:19:31 crc kubenswrapper[4967]: I1011 04:19:31.724168 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_c41a3ead-949a-4f76-9a58-0e1b4db0abb5/rabbitmq/0.log" Oct 11 04:19:31 crc kubenswrapper[4967]: I1011 04:19:31.860818 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab/setup-container/0.log" Oct 11 04:19:31 crc kubenswrapper[4967]: I1011 04:19:31.878131 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab/rabbitmq/0.log" Oct 11 04:19:31 crc kubenswrapper[4967]: I1011 04:19:31.995242 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6486584599-dfqxj_fe5d2413-57a2-449d-9973-642e00e8c49f/proxy-httpd/0.log" Oct 11 04:19:32 crc kubenswrapper[4967]: I1011 04:19:32.089496 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6486584599-dfqxj_fe5d2413-57a2-449d-9973-642e00e8c49f/proxy-server/0.log" Oct 11 04:19:32 crc kubenswrapper[4967]: I1011 04:19:32.166426 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-6rr4l_3f2085d6-0a42-4e1e-b52d-e9379d2a4838/swift-ring-rebalance/0.log" Oct 11 04:19:32 crc kubenswrapper[4967]: I1011 04:19:32.317348 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c09e3333-ad4d-495f-8973-7e2d2f0931d4/account-auditor/0.log" Oct 11 04:19:32 crc kubenswrapper[4967]: I1011 04:19:32.350676 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c09e3333-ad4d-495f-8973-7e2d2f0931d4/account-reaper/0.log" Oct 11 04:19:32 crc kubenswrapper[4967]: I1011 04:19:32.434666 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c09e3333-ad4d-495f-8973-7e2d2f0931d4/account-replicator/0.log" Oct 11 04:19:32 crc kubenswrapper[4967]: I1011 04:19:32.437658 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c09e3333-ad4d-495f-8973-7e2d2f0931d4/account-server/0.log" Oct 11 04:19:32 crc kubenswrapper[4967]: I1011 04:19:32.592886 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c09e3333-ad4d-495f-8973-7e2d2f0931d4/container-replicator/0.log" Oct 11 04:19:32 crc kubenswrapper[4967]: I1011 04:19:32.601980 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c09e3333-ad4d-495f-8973-7e2d2f0931d4/container-auditor/0.log" Oct 11 04:19:32 crc kubenswrapper[4967]: I1011 04:19:32.686647 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c09e3333-ad4d-495f-8973-7e2d2f0931d4/container-server/0.log" Oct 11 04:19:32 crc kubenswrapper[4967]: I1011 04:19:32.705567 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c09e3333-ad4d-495f-8973-7e2d2f0931d4/container-updater/0.log" Oct 11 04:19:32 crc kubenswrapper[4967]: I1011 04:19:32.833211 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c09e3333-ad4d-495f-8973-7e2d2f0931d4/object-auditor/0.log" Oct 11 04:19:32 crc kubenswrapper[4967]: I1011 04:19:32.894952 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c09e3333-ad4d-495f-8973-7e2d2f0931d4/object-replicator/0.log" Oct 11 04:19:32 crc kubenswrapper[4967]: I1011 04:19:32.907704 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c09e3333-ad4d-495f-8973-7e2d2f0931d4/object-expirer/0.log" Oct 11 04:19:32 crc kubenswrapper[4967]: I1011 04:19:32.920020 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c09e3333-ad4d-495f-8973-7e2d2f0931d4/object-server/0.log" Oct 11 04:19:33 crc kubenswrapper[4967]: I1011 04:19:33.083292 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c09e3333-ad4d-495f-8973-7e2d2f0931d4/rsync/0.log" Oct 11 04:19:33 crc kubenswrapper[4967]: I1011 04:19:33.119614 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c09e3333-ad4d-495f-8973-7e2d2f0931d4/swift-recon-cron/0.log" Oct 11 04:19:33 crc kubenswrapper[4967]: I1011 04:19:33.124807 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c09e3333-ad4d-495f-8973-7e2d2f0931d4/object-updater/0.log" Oct 11 04:19:34 crc kubenswrapper[4967]: I1011 04:19:34.227490 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_10ea033f-281c-4519-8da7-93e598f08b11/memcached/0.log" Oct 11 04:19:37 crc kubenswrapper[4967]: I1011 04:19:37.045967 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xh6jw"] Oct 11 04:19:37 crc kubenswrapper[4967]: I1011 04:19:37.057966 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-conductor-db-sync-xh6jw"] Oct 11 04:19:38 crc kubenswrapper[4967]: I1011 04:19:38.826479 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="d4a234ae-b210-4ac3-9bf2-841f586ee4a3" path="/var/lib/kubelet/pods/d4a234ae-b210-4ac3-9bf2-841f586ee4a3/volumes" Oct 11 04:19:39 crc kubenswrapper[4967]: I1011 04:19:39.815427 4967 scope.go:117] "RemoveContainer" containerID="6010a3358387442fdc0ff1dfa8bf2d3fb285ef11cfb5f18ad38fe86548ca1f20" Oct 11 04:19:39 crc kubenswrapper[4967]: E1011 04:19:39.816172 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:19:50 crc kubenswrapper[4967]: I1011 04:19:50.816028 4967 scope.go:117] "RemoveContainer" containerID="6010a3358387442fdc0ff1dfa8bf2d3fb285ef11cfb5f18ad38fe86548ca1f20" Oct 11 04:19:50 crc kubenswrapper[4967]: E1011 04:19:50.817004 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:19:54 crc kubenswrapper[4967]: I1011 04:19:54.601060 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk_9adf77e3-238d-4024-99db-ab21bdd51110/util/0.log" Oct 11 04:19:54 crc kubenswrapper[4967]: I1011 04:19:54.729820 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk_9adf77e3-238d-4024-99db-ab21bdd51110/util/0.log" Oct 11 04:19:54 crc kubenswrapper[4967]: I1011 04:19:54.730947 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk_9adf77e3-238d-4024-99db-ab21bdd51110/pull/0.log" Oct 11 04:19:54 crc kubenswrapper[4967]: I1011 04:19:54.796983 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk_9adf77e3-238d-4024-99db-ab21bdd51110/pull/0.log" Oct 11 04:19:54 crc kubenswrapper[4967]: I1011 04:19:54.946123 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk_9adf77e3-238d-4024-99db-ab21bdd51110/util/0.log" Oct 11 04:19:54 crc kubenswrapper[4967]: I1011 04:19:54.992917 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk_9adf77e3-238d-4024-99db-ab21bdd51110/pull/0.log" Oct 11 04:19:54 crc kubenswrapper[4967]: I1011 04:19:54.997862 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk_9adf77e3-238d-4024-99db-ab21bdd51110/extract/0.log" Oct 11 04:19:55 crc kubenswrapper[4967]: I1011 04:19:55.136936 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-xz92m_6f3fbb7d-1207-45a2-8213-c7c71bd703f1/kube-rbac-proxy/0.log" Oct 11 04:19:55 crc kubenswrapper[4967]: I1011 04:19:55.212427 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-xz92m_6f3fbb7d-1207-45a2-8213-c7c71bd703f1/manager/0.log" Oct 11 04:19:55 crc kubenswrapper[4967]: I1011 04:19:55.233768 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-5glgp_91f45294-2d2f-4b40-a1b4-a841c7748cc8/kube-rbac-proxy/0.log" Oct 11 04:19:55 crc kubenswrapper[4967]: I1011 04:19:55.340492 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-5glgp_91f45294-2d2f-4b40-a1b4-a841c7748cc8/manager/0.log" Oct 11 04:19:55 crc kubenswrapper[4967]: I1011 04:19:55.388065 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-9gn9l_2a2805ea-7d92-42e4-89cb-9b2672f72fc6/kube-rbac-proxy/0.log" Oct 11 04:19:55 crc kubenswrapper[4967]: I1011 04:19:55.410336 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-9gn9l_2a2805ea-7d92-42e4-89cb-9b2672f72fc6/manager/0.log" Oct 11 04:19:55 crc kubenswrapper[4967]: I1011 04:19:55.533064 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-zt44b_44eb39fa-5b20-4c02-b039-f23d7c5220d2/kube-rbac-proxy/0.log" Oct 11 04:19:55 crc kubenswrapper[4967]: I1011 04:19:55.627695 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-zt44b_44eb39fa-5b20-4c02-b039-f23d7c5220d2/manager/0.log" Oct 11 04:19:55 crc kubenswrapper[4967]: I1011 04:19:55.711270 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-ww7nt_ecd905f3-adb2-47ae-a86d-6fc0b4907a76/manager/0.log" Oct 11 04:19:55 crc kubenswrapper[4967]: I1011 04:19:55.752046 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-ww7nt_ecd905f3-adb2-47ae-a86d-6fc0b4907a76/kube-rbac-proxy/0.log" Oct 11 04:19:55 crc kubenswrapper[4967]: I1011 04:19:55.794089 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-8ccrl_e9feb54d-dbfa-40ec-ac45-535cca431024/kube-rbac-proxy/0.log" Oct 11 04:19:55 crc kubenswrapper[4967]: I1011 04:19:55.872761 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-8ccrl_e9feb54d-dbfa-40ec-ac45-535cca431024/manager/0.log" Oct 11 04:19:55 crc kubenswrapper[4967]: I1011 04:19:55.972193 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-685c45897-5gbwb_6628a0ae-2a55-48b0-a5f1-0439c7d0c273/kube-rbac-proxy/0.log" Oct 11 04:19:56 crc kubenswrapper[4967]: I1011 04:19:56.123592 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67494cb947-sklz6_674852f1-9729-4a6e-8c90-31741fa69b49/kube-rbac-proxy/0.log" Oct 11 04:19:56 crc kubenswrapper[4967]: I1011 04:19:56.148538 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-685c45897-5gbwb_6628a0ae-2a55-48b0-a5f1-0439c7d0c273/manager/0.log" Oct 11 04:19:56 crc kubenswrapper[4967]: I1011 04:19:56.239139 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67494cb947-sklz6_674852f1-9729-4a6e-8c90-31741fa69b49/manager/0.log" Oct 11 04:19:56 crc kubenswrapper[4967]: I1011 04:19:56.323376 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-ddb98f99b-brbrs_950d3635-f160-45a1-9a4c-749720e662dc/kube-rbac-proxy/0.log" Oct 11 04:19:56 crc kubenswrapper[4967]: I1011 04:19:56.408720 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-ddb98f99b-brbrs_950d3635-f160-45a1-9a4c-749720e662dc/manager/0.log" Oct 11 04:19:56 crc kubenswrapper[4967]: I1011 04:19:56.484749 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-ls9k6_20584c10-d8a1-41ca-a0a2-9465fa5190f2/kube-rbac-proxy/0.log" Oct 11 04:19:56 crc kubenswrapper[4967]: I1011 04:19:56.518035 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-ls9k6_20584c10-d8a1-41ca-a0a2-9465fa5190f2/manager/0.log" Oct 11 04:19:56 crc kubenswrapper[4967]: I1011 04:19:56.599689 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-fjr8k_ef8ae8f4-ad95-4f1f-bc8c-170fe5058dc9/kube-rbac-proxy/0.log" Oct 11 04:19:56 crc kubenswrapper[4967]: I1011 04:19:56.684195 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-fjr8k_ef8ae8f4-ad95-4f1f-bc8c-170fe5058dc9/manager/0.log" Oct 11 04:19:56 crc kubenswrapper[4967]: I1011 04:19:56.785020 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-6zmv5_538ebbcb-d51c-472c-a2b4-80236b0c8c08/kube-rbac-proxy/0.log" Oct 11 04:19:56 crc kubenswrapper[4967]: I1011 04:19:56.861780 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-6zmv5_538ebbcb-d51c-472c-a2b4-80236b0c8c08/manager/0.log" Oct 11 04:19:56 crc kubenswrapper[4967]: I1011 04:19:56.890151 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-59px8_7140d0e1-9937-491b-aa8a-16db830aff11/kube-rbac-proxy/0.log" Oct 11 04:19:57 crc kubenswrapper[4967]: I1011 04:19:57.040646 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-59px8_7140d0e1-9937-491b-aa8a-16db830aff11/manager/0.log" Oct 11 04:19:57 crc kubenswrapper[4967]: I1011 04:19:57.069891 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-7clg2_7e917a6a-9720-4427-b079-33d112a981a3/manager/0.log" Oct 11 04:19:57 crc kubenswrapper[4967]: I1011 04:19:57.104566 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-7clg2_7e917a6a-9720-4427-b079-33d112a981a3/kube-rbac-proxy/0.log" Oct 11 04:19:57 crc kubenswrapper[4967]: I1011 04:19:57.231473 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns_0cc11444-d341-4821-a743-6b4af1466e4a/kube-rbac-proxy/0.log" Oct 11 04:19:57 crc kubenswrapper[4967]: I1011 04:19:57.240120 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns_0cc11444-d341-4821-a743-6b4af1466e4a/manager/0.log" Oct 11 04:19:57 crc kubenswrapper[4967]: I1011 04:19:57.406269 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-66664b49c9-tk2lg_620877eb-6f9c-4647-b545-1cca7a438690/kube-rbac-proxy/0.log" Oct 11 04:19:57 crc kubenswrapper[4967]: I1011 04:19:57.526139 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-84859cd6c-hlmz6_e36d39c8-9b9a-4564-b779-0c038ba68dfb/kube-rbac-proxy/0.log" Oct 11 04:19:57 crc kubenswrapper[4967]: I1011 04:19:57.739641 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-xvswq_3dafc1e8-5bc6-4b6b-baa4-8c8204d98d31/registry-server/0.log" Oct 11 04:19:57 crc kubenswrapper[4967]: I1011 04:19:57.903264 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-84859cd6c-hlmz6_e36d39c8-9b9a-4564-b779-0c038ba68dfb/operator/0.log" Oct 11 04:19:57 crc kubenswrapper[4967]: I1011 04:19:57.946002 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-869cc7797f-kkrkc_c9e29e65-cdd1-408d-bed0-bc283147b283/kube-rbac-proxy/0.log" Oct 11 04:19:57 crc kubenswrapper[4967]: I1011 04:19:57.980891 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-869cc7797f-kkrkc_c9e29e65-cdd1-408d-bed0-bc283147b283/manager/0.log" Oct 11 04:19:58 crc kubenswrapper[4967]: I1011 04:19:58.135055 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-vng4c_26b9d5a0-dfbe-4e53-aa03-4fbb4682019b/kube-rbac-proxy/0.log" Oct 11 04:19:58 crc kubenswrapper[4967]: I1011 04:19:58.223782 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-vng4c_26b9d5a0-dfbe-4e53-aa03-4fbb4682019b/manager/0.log" Oct 11 04:19:58 crc kubenswrapper[4967]: I1011 04:19:58.258302 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-66664b49c9-tk2lg_620877eb-6f9c-4647-b545-1cca7a438690/manager/0.log" Oct 11 04:19:58 crc kubenswrapper[4967]: I1011 04:19:58.285711 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-cmzk2_11e4f97c-6107-45fd-8eb1-e39ef7bed4ce/operator/0.log" Oct 11 04:19:58 crc kubenswrapper[4967]: I1011 04:19:58.435892 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-65knj_1094a8be-16bb-4986-b4a4-465d41880800/manager/0.log" Oct 11 04:19:58 crc kubenswrapper[4967]: I1011 04:19:58.436747 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-65knj_1094a8be-16bb-4986-b4a4-465d41880800/kube-rbac-proxy/0.log" Oct 11 04:19:58 crc kubenswrapper[4967]: I1011 04:19:58.477279 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-578874c84d-fgz82_e82102ad-391b-4b29-ae47-dae18dba7743/kube-rbac-proxy/0.log" Oct 11 04:19:58 crc kubenswrapper[4967]: I1011 04:19:58.615703 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-578874c84d-fgz82_e82102ad-391b-4b29-ae47-dae18dba7743/manager/0.log" Oct 11 04:19:58 crc kubenswrapper[4967]: I1011 04:19:58.641915 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-ffcdd6c94-4s7hv_5e24d70e-8882-45c0-9f6b-e443549ecabb/kube-rbac-proxy/0.log" Oct 11 04:19:58 crc kubenswrapper[4967]: I1011 04:19:58.692579 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-ffcdd6c94-4s7hv_5e24d70e-8882-45c0-9f6b-e443549ecabb/manager/0.log" Oct 11 04:19:58 crc kubenswrapper[4967]: I1011 04:19:58.787143 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-646675d848-zdpbn_10a0ea36-9248-4086-83e4-de8578492421/kube-rbac-proxy/0.log" Oct 11 04:19:58 crc kubenswrapper[4967]: I1011 04:19:58.817371 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-646675d848-zdpbn_10a0ea36-9248-4086-83e4-de8578492421/manager/0.log" Oct 11 04:20:01 crc kubenswrapper[4967]: I1011 04:20:01.035736 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell0-cell-mapping-krtvp"] Oct 11 04:20:01 crc kubenswrapper[4967]: I1011 04:20:01.046574 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell0-cell-mapping-krtvp"] Oct 11 04:20:02 crc kubenswrapper[4967]: I1011 04:20:02.824974 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e" path="/var/lib/kubelet/pods/9f4b6d1f-5bf1-4784-aaa3-0c4b5b6a224e/volumes" Oct 11 04:20:05 crc kubenswrapper[4967]: I1011 04:20:05.815917 4967 scope.go:117] "RemoveContainer" containerID="6010a3358387442fdc0ff1dfa8bf2d3fb285ef11cfb5f18ad38fe86548ca1f20" Oct 11 04:20:05 crc kubenswrapper[4967]: E1011 04:20:05.816516 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:20:06 crc kubenswrapper[4967]: I1011 04:20:06.035686 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dt5p8"] Oct 11 04:20:06 crc kubenswrapper[4967]: I1011 04:20:06.047059 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-conductor-db-sync-dt5p8"] Oct 11 04:20:06 crc kubenswrapper[4967]: I1011 04:20:06.829729 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="702714d0-d935-49ae-b19f-c0f80acd7758" path="/var/lib/kubelet/pods/702714d0-d935-49ae-b19f-c0f80acd7758/volumes" Oct 11 04:20:14 crc kubenswrapper[4967]: I1011 04:20:14.334889 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-rv6n2_c3a20ddb-cfaa-46ac-8aed-4db9de3e32b6/control-plane-machine-set-operator/0.log" Oct 11 04:20:14 crc kubenswrapper[4967]: I1011 04:20:14.519034 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-mrsjk_9b3efa80-8370-4353-b688-4701c4deec8f/kube-rbac-proxy/0.log" Oct 11 04:20:14 crc kubenswrapper[4967]: I1011 04:20:14.535925 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-mrsjk_9b3efa80-8370-4353-b688-4701c4deec8f/machine-api-operator/0.log" Oct 11 04:20:17 crc kubenswrapper[4967]: I1011 04:20:17.815282 4967 scope.go:117] "RemoveContainer" containerID="6010a3358387442fdc0ff1dfa8bf2d3fb285ef11cfb5f18ad38fe86548ca1f20" Oct 11 04:20:17 crc kubenswrapper[4967]: E1011 04:20:17.815873 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:20:25 crc kubenswrapper[4967]: I1011 04:20:25.371781 4967 scope.go:117] "RemoveContainer" containerID="8bdf9296be5a3e6cedf0cf960a89608cbc5c90a2175a407f8d332f1f82d80cc8" Oct 11 04:20:25 crc kubenswrapper[4967]: I1011 04:20:25.414729 4967 scope.go:117] "RemoveContainer" containerID="4bbf287d1de495b499f6b4194a0015fc0ee50f432089740f8c0298235e83628f" Oct 11 04:20:25 crc kubenswrapper[4967]: I1011 04:20:25.468025 4967 scope.go:117] "RemoveContainer" containerID="a31acefb52ee3f66f9f084ee9422d5cd5026dded84e8d4de97d3d86de40ab227" Oct 11 04:20:27 crc kubenswrapper[4967]: I1011 04:20:27.010448 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-ml5f9_be6fdb4b-abb9-4e76-be7e-0c7c579d3bf8/cert-manager-controller/0.log" Oct 11 04:20:27 crc kubenswrapper[4967]: I1011 04:20:27.132682 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-ncqxv_822cb861-f537-4669-8fc9-168f052a4b5d/cert-manager-cainjector/0.log" Oct 11 04:20:27 crc kubenswrapper[4967]: I1011 04:20:27.150042 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-xmk8c_ac43a6d2-5018-4d03-9a26-5da4f2619403/cert-manager-webhook/0.log" Oct 11 04:20:30 crc kubenswrapper[4967]: I1011 04:20:30.815055 4967 scope.go:117] "RemoveContainer" containerID="6010a3358387442fdc0ff1dfa8bf2d3fb285ef11cfb5f18ad38fe86548ca1f20" Oct 11 04:20:30 crc kubenswrapper[4967]: E1011 04:20:30.816005 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:20:39 crc kubenswrapper[4967]: I1011 04:20:39.361904 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-kfkw5_215e47d4-e289-4a93-91a7-f64b901bebeb/nmstate-console-plugin/0.log" Oct 11 04:20:39 crc kubenswrapper[4967]: I1011 04:20:39.553644 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-6qvqt_b5c62540-5e0d-4601-9d0b-bef391800fe7/nmstate-handler/0.log" Oct 11 04:20:39 crc kubenswrapper[4967]: I1011 04:20:39.569873 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-bhrz7_abab15af-313d-42ef-96ce-4c0507429f93/kube-rbac-proxy/0.log" Oct 11 04:20:39 crc kubenswrapper[4967]: I1011 04:20:39.609227 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-bhrz7_abab15af-313d-42ef-96ce-4c0507429f93/nmstate-metrics/0.log" Oct 11 04:20:39 crc kubenswrapper[4967]: I1011 04:20:39.752836 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-x6r8r_419acadb-11b8-4e90-83f8-b02a39264a04/nmstate-operator/0.log" Oct 11 04:20:39 crc kubenswrapper[4967]: I1011 04:20:39.819933 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-4lvvm_c23b1485-a6f5-4504-bc81-a069d2a96182/nmstate-webhook/0.log" Oct 11 04:20:44 crc kubenswrapper[4967]: I1011 04:20:44.814855 4967 scope.go:117] "RemoveContainer" containerID="6010a3358387442fdc0ff1dfa8bf2d3fb285ef11cfb5f18ad38fe86548ca1f20" Oct 11 04:20:44 crc kubenswrapper[4967]: E1011 04:20:44.815840 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:20:50 crc kubenswrapper[4967]: I1011 04:20:50.055285 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openstack/nova-cell1-cell-mapping-2nl8n"] Oct 11 04:20:50 crc kubenswrapper[4967]: I1011 04:20:50.064564 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openstack/nova-cell1-cell-mapping-2nl8n"] Oct 11 04:20:50 crc kubenswrapper[4967]: I1011 04:20:50.832561 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="11ef0366-73db-43a2-88c4-780d2a0084c9" path="/var/lib/kubelet/pods/11ef0366-73db-43a2-88c4-780d2a0084c9/volumes" Oct 11 04:20:54 crc kubenswrapper[4967]: I1011 04:20:54.807951 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-rv747_7fb0bb1d-4f24-4ca2-8148-ac03ee0083c5/kube-rbac-proxy/0.log" Oct 11 04:20:54 crc kubenswrapper[4967]: I1011 04:20:54.998407 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-rv747_7fb0bb1d-4f24-4ca2-8148-ac03ee0083c5/controller/0.log" Oct 11 04:20:55 crc kubenswrapper[4967]: I1011 04:20:55.047905 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/cp-frr-files/0.log" Oct 11 04:20:55 crc kubenswrapper[4967]: I1011 04:20:55.253525 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/cp-frr-files/0.log" Oct 11 04:20:55 crc kubenswrapper[4967]: I1011 04:20:55.259066 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/cp-reloader/0.log" Oct 11 04:20:55 crc kubenswrapper[4967]: I1011 04:20:55.300652 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/cp-metrics/0.log" Oct 11 04:20:55 crc kubenswrapper[4967]: I1011 04:20:55.301969 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/cp-reloader/0.log" Oct 11 04:20:55 crc kubenswrapper[4967]: I1011 04:20:55.441451 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/cp-reloader/0.log" Oct 11 04:20:55 crc kubenswrapper[4967]: I1011 04:20:55.456443 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/cp-frr-files/0.log" Oct 11 04:20:55 crc kubenswrapper[4967]: I1011 04:20:55.494946 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/cp-metrics/0.log" Oct 11 04:20:55 crc kubenswrapper[4967]: I1011 04:20:55.502185 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/cp-metrics/0.log" Oct 11 04:20:55 crc kubenswrapper[4967]: I1011 04:20:55.654380 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/cp-reloader/0.log" Oct 11 04:20:55 crc kubenswrapper[4967]: I1011 04:20:55.676589 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/cp-metrics/0.log" Oct 11 04:20:55 crc kubenswrapper[4967]: I1011 04:20:55.680290 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/controller/0.log" Oct 11 04:20:55 crc kubenswrapper[4967]: I1011 04:20:55.688048 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/cp-frr-files/0.log" Oct 11 04:20:55 crc kubenswrapper[4967]: I1011 04:20:55.843374 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/kube-rbac-proxy-frr/0.log" Oct 11 04:20:55 crc kubenswrapper[4967]: I1011 04:20:55.852756 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/frr-metrics/0.log" Oct 11 04:20:55 crc kubenswrapper[4967]: I1011 04:20:55.911396 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/kube-rbac-proxy/0.log" Oct 11 04:20:56 crc kubenswrapper[4967]: I1011 04:20:56.082755 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/reloader/0.log" Oct 11 04:20:56 crc kubenswrapper[4967]: I1011 04:20:56.149624 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-j9lg6_50110bd6-5363-43c8-b3a8-b4b3762c0c43/frr-k8s-webhook-server/0.log" Oct 11 04:20:56 crc kubenswrapper[4967]: I1011 04:20:56.357050 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5cfc444676-2s7hf_f3173c9b-89b4-420d-a852-f293d581da52/manager/0.log" Oct 11 04:20:56 crc kubenswrapper[4967]: I1011 04:20:56.536919 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-56f68cd588-8fj48_c88642bc-96a7-4662-a035-c2c6043e322f/webhook-server/0.log" Oct 11 04:20:56 crc kubenswrapper[4967]: I1011 04:20:56.552494 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-np2lm_816b70d7-ff40-4cee-b21d-01008f0d4868/kube-rbac-proxy/0.log" Oct 11 04:20:56 crc kubenswrapper[4967]: I1011 04:20:56.843942 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/frr/0.log" Oct 11 04:20:57 crc kubenswrapper[4967]: I1011 04:20:57.019940 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-np2lm_816b70d7-ff40-4cee-b21d-01008f0d4868/speaker/0.log" Oct 11 04:20:59 crc kubenswrapper[4967]: I1011 04:20:59.814892 4967 scope.go:117] "RemoveContainer" containerID="6010a3358387442fdc0ff1dfa8bf2d3fb285ef11cfb5f18ad38fe86548ca1f20" Oct 11 04:20:59 crc kubenswrapper[4967]: E1011 04:20:59.815567 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:21:09 crc kubenswrapper[4967]: I1011 04:21:09.232431 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn_bd662d9c-6ab0-4518-8fab-83115ccd3569/util/0.log" Oct 11 04:21:09 crc kubenswrapper[4967]: I1011 04:21:09.375504 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn_bd662d9c-6ab0-4518-8fab-83115ccd3569/util/0.log" Oct 11 04:21:09 crc kubenswrapper[4967]: I1011 04:21:09.385732 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn_bd662d9c-6ab0-4518-8fab-83115ccd3569/pull/0.log" Oct 11 04:21:09 crc kubenswrapper[4967]: I1011 04:21:09.406915 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn_bd662d9c-6ab0-4518-8fab-83115ccd3569/pull/0.log" Oct 11 04:21:09 crc kubenswrapper[4967]: I1011 04:21:09.580767 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn_bd662d9c-6ab0-4518-8fab-83115ccd3569/util/0.log" Oct 11 04:21:09 crc kubenswrapper[4967]: I1011 04:21:09.581575 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn_bd662d9c-6ab0-4518-8fab-83115ccd3569/pull/0.log" Oct 11 04:21:09 crc kubenswrapper[4967]: I1011 04:21:09.642519 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn_bd662d9c-6ab0-4518-8fab-83115ccd3569/extract/0.log" Oct 11 04:21:09 crc kubenswrapper[4967]: I1011 04:21:09.802413 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mxkbg_afb60be7-f83c-4dc3-9129-ec66a8b75a55/extract-utilities/0.log" Oct 11 04:21:09 crc kubenswrapper[4967]: I1011 04:21:09.923824 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mxkbg_afb60be7-f83c-4dc3-9129-ec66a8b75a55/extract-utilities/0.log" Oct 11 04:21:09 crc kubenswrapper[4967]: I1011 04:21:09.925863 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mxkbg_afb60be7-f83c-4dc3-9129-ec66a8b75a55/extract-content/0.log" Oct 11 04:21:09 crc kubenswrapper[4967]: I1011 04:21:09.950731 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mxkbg_afb60be7-f83c-4dc3-9129-ec66a8b75a55/extract-content/0.log" Oct 11 04:21:10 crc kubenswrapper[4967]: I1011 04:21:10.082967 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mxkbg_afb60be7-f83c-4dc3-9129-ec66a8b75a55/extract-utilities/0.log" Oct 11 04:21:10 crc kubenswrapper[4967]: I1011 04:21:10.127952 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mxkbg_afb60be7-f83c-4dc3-9129-ec66a8b75a55/extract-content/0.log" Oct 11 04:21:10 crc kubenswrapper[4967]: I1011 04:21:10.348038 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dgst6_cd79944d-93c5-45d4-b68d-5cff92c3d111/extract-utilities/0.log" Oct 11 04:21:10 crc kubenswrapper[4967]: I1011 04:21:10.359834 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mxkbg_afb60be7-f83c-4dc3-9129-ec66a8b75a55/registry-server/0.log" Oct 11 04:21:10 crc kubenswrapper[4967]: I1011 04:21:10.535041 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dgst6_cd79944d-93c5-45d4-b68d-5cff92c3d111/extract-utilities/0.log" Oct 11 04:21:10 crc kubenswrapper[4967]: I1011 04:21:10.539175 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dgst6_cd79944d-93c5-45d4-b68d-5cff92c3d111/extract-content/0.log" Oct 11 04:21:10 crc kubenswrapper[4967]: I1011 04:21:10.553660 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dgst6_cd79944d-93c5-45d4-b68d-5cff92c3d111/extract-content/0.log" Oct 11 04:21:10 crc kubenswrapper[4967]: I1011 04:21:10.704104 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dgst6_cd79944d-93c5-45d4-b68d-5cff92c3d111/extract-content/0.log" Oct 11 04:21:10 crc kubenswrapper[4967]: I1011 04:21:10.723812 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dgst6_cd79944d-93c5-45d4-b68d-5cff92c3d111/extract-utilities/0.log" Oct 11 04:21:10 crc kubenswrapper[4967]: I1011 04:21:10.819977 4967 scope.go:117] "RemoveContainer" containerID="6010a3358387442fdc0ff1dfa8bf2d3fb285ef11cfb5f18ad38fe86548ca1f20" Oct 11 04:21:10 crc kubenswrapper[4967]: E1011 04:21:10.820223 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:21:10 crc kubenswrapper[4967]: I1011 04:21:10.924691 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6_afec4044-82ab-4b5b-8669-7605ab680e10/util/0.log" Oct 11 04:21:10 crc kubenswrapper[4967]: I1011 04:21:10.985917 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dgst6_cd79944d-93c5-45d4-b68d-5cff92c3d111/registry-server/0.log" Oct 11 04:21:11 crc kubenswrapper[4967]: I1011 04:21:11.097346 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6_afec4044-82ab-4b5b-8669-7605ab680e10/util/0.log" Oct 11 04:21:11 crc kubenswrapper[4967]: I1011 04:21:11.138418 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6_afec4044-82ab-4b5b-8669-7605ab680e10/pull/0.log" Oct 11 04:21:11 crc kubenswrapper[4967]: I1011 04:21:11.151782 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6_afec4044-82ab-4b5b-8669-7605ab680e10/pull/0.log" Oct 11 04:21:11 crc kubenswrapper[4967]: I1011 04:21:11.327217 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6_afec4044-82ab-4b5b-8669-7605ab680e10/util/0.log" Oct 11 04:21:11 crc kubenswrapper[4967]: I1011 04:21:11.343041 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6_afec4044-82ab-4b5b-8669-7605ab680e10/pull/0.log" Oct 11 04:21:11 crc kubenswrapper[4967]: I1011 04:21:11.348040 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6_afec4044-82ab-4b5b-8669-7605ab680e10/extract/0.log" Oct 11 04:21:11 crc kubenswrapper[4967]: I1011 04:21:11.489865 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-sp7h8_020dc4f8-c49b-4447-8373-426bf747cace/marketplace-operator/0.log" Oct 11 04:21:11 crc kubenswrapper[4967]: I1011 04:21:11.505235 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p2xpx_065a2523-459f-440e-b5da-5ea5318fd89f/extract-utilities/0.log" Oct 11 04:21:11 crc kubenswrapper[4967]: I1011 04:21:11.672621 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p2xpx_065a2523-459f-440e-b5da-5ea5318fd89f/extract-content/0.log" Oct 11 04:21:11 crc kubenswrapper[4967]: I1011 04:21:11.673251 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p2xpx_065a2523-459f-440e-b5da-5ea5318fd89f/extract-utilities/0.log" Oct 11 04:21:11 crc kubenswrapper[4967]: I1011 04:21:11.693387 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p2xpx_065a2523-459f-440e-b5da-5ea5318fd89f/extract-content/0.log" Oct 11 04:21:11 crc kubenswrapper[4967]: I1011 04:21:11.874670 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p2xpx_065a2523-459f-440e-b5da-5ea5318fd89f/extract-content/0.log" Oct 11 04:21:11 crc kubenswrapper[4967]: I1011 04:21:11.903458 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p2xpx_065a2523-459f-440e-b5da-5ea5318fd89f/extract-utilities/0.log" Oct 11 04:21:11 crc kubenswrapper[4967]: I1011 04:21:11.912391 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p2xpx_065a2523-459f-440e-b5da-5ea5318fd89f/registry-server/0.log" Oct 11 04:21:12 crc kubenswrapper[4967]: I1011 04:21:12.055630 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-khtzc_553fcde4-86f4-4a03-9650-bb6bd3ee5d0e/extract-utilities/0.log" Oct 11 04:21:12 crc kubenswrapper[4967]: I1011 04:21:12.202294 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-khtzc_553fcde4-86f4-4a03-9650-bb6bd3ee5d0e/extract-utilities/0.log" Oct 11 04:21:12 crc kubenswrapper[4967]: I1011 04:21:12.243016 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-khtzc_553fcde4-86f4-4a03-9650-bb6bd3ee5d0e/extract-content/0.log" Oct 11 04:21:12 crc kubenswrapper[4967]: I1011 04:21:12.247587 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-khtzc_553fcde4-86f4-4a03-9650-bb6bd3ee5d0e/extract-content/0.log" Oct 11 04:21:12 crc kubenswrapper[4967]: I1011 04:21:12.387407 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-khtzc_553fcde4-86f4-4a03-9650-bb6bd3ee5d0e/extract-content/0.log" Oct 11 04:21:12 crc kubenswrapper[4967]: I1011 04:21:12.393808 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-khtzc_553fcde4-86f4-4a03-9650-bb6bd3ee5d0e/extract-utilities/0.log" Oct 11 04:21:12 crc kubenswrapper[4967]: I1011 04:21:12.640869 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-khtzc_553fcde4-86f4-4a03-9650-bb6bd3ee5d0e/registry-server/0.log" Oct 11 04:21:24 crc kubenswrapper[4967]: I1011 04:21:24.815824 4967 scope.go:117] "RemoveContainer" containerID="6010a3358387442fdc0ff1dfa8bf2d3fb285ef11cfb5f18ad38fe86548ca1f20" Oct 11 04:21:25 crc kubenswrapper[4967]: I1011 04:21:25.600658 4967 scope.go:117] "RemoveContainer" containerID="910da87b7c970cfcc4b89b941d30c32a87f174c3e567e6f92be4f17019eb4749" Oct 11 04:21:25 crc kubenswrapper[4967]: I1011 04:21:25.913469 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" event={"ID":"97c07678-14be-410c-b61f-498cb49bc960","Type":"ContainerStarted","Data":"d6d902b7233db28561ab70d9cf62192d380aef02bd4e78fc79f31d3c320df63c"} Oct 11 04:21:43 crc kubenswrapper[4967]: E1011 04:21:43.213976 4967 upgradeaware.go:441] Error proxying data from backend to client: writeto tcp 38.102.83.50:35322->38.102.83.50:40947: read tcp 38.102.83.50:35322->38.102.83.50:40947: read: connection reset by peer Oct 11 04:21:47 crc kubenswrapper[4967]: E1011 04:21:47.389200 4967 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.50:50922->38.102.83.50:40947: write tcp 38.102.83.50:50922->38.102.83.50:40947: write: broken pipe Oct 11 04:22:41 crc kubenswrapper[4967]: I1011 04:22:41.745511 4967 generic.go:334] "Generic (PLEG): container finished" podID="31a11147-0dcb-4957-826d-6579699bde94" containerID="f672d23f2106ef666680d24be5d2207af3059c78ee3b47c6c60c33f34e34e223" exitCode=0 Oct 11 04:22:41 crc kubenswrapper[4967]: I1011 04:22:41.745574 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-vt6vn/must-gather-689h5" event={"ID":"31a11147-0dcb-4957-826d-6579699bde94","Type":"ContainerDied","Data":"f672d23f2106ef666680d24be5d2207af3059c78ee3b47c6c60c33f34e34e223"} Oct 11 04:22:41 crc kubenswrapper[4967]: I1011 04:22:41.750629 4967 scope.go:117] "RemoveContainer" containerID="f672d23f2106ef666680d24be5d2207af3059c78ee3b47c6c60c33f34e34e223" Oct 11 04:22:41 crc kubenswrapper[4967]: I1011 04:22:41.841522 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-vt6vn_must-gather-689h5_31a11147-0dcb-4957-826d-6579699bde94/gather/0.log" Oct 11 04:22:48 crc kubenswrapper[4967]: I1011 04:22:48.712161 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-vt6vn/must-gather-689h5"] Oct 11 04:22:48 crc kubenswrapper[4967]: I1011 04:22:48.712962 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-vt6vn/must-gather-689h5" podUID="31a11147-0dcb-4957-826d-6579699bde94" containerName="copy" containerID="cri-o://f6b72a8eac803d353e976a61f4ebfb8f5c2e6ad41329bb8a50f24479247338a7" gracePeriod=2 Oct 11 04:22:48 crc kubenswrapper[4967]: I1011 04:22:48.726032 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-vt6vn/must-gather-689h5"] Oct 11 04:22:49 crc kubenswrapper[4967]: I1011 04:22:49.171727 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-vt6vn_must-gather-689h5_31a11147-0dcb-4957-826d-6579699bde94/copy/0.log" Oct 11 04:22:49 crc kubenswrapper[4967]: I1011 04:22:49.172671 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vt6vn/must-gather-689h5" Oct 11 04:22:49 crc kubenswrapper[4967]: I1011 04:22:49.338971 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-hxxgw\" (UniqueName: \"kubernetes.io/projected/31a11147-0dcb-4957-826d-6579699bde94-kube-api-access-hxxgw\") pod \"31a11147-0dcb-4957-826d-6579699bde94\" (UID: \"31a11147-0dcb-4957-826d-6579699bde94\") " Oct 11 04:22:49 crc kubenswrapper[4967]: I1011 04:22:49.339223 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/31a11147-0dcb-4957-826d-6579699bde94-must-gather-output\") pod \"31a11147-0dcb-4957-826d-6579699bde94\" (UID: \"31a11147-0dcb-4957-826d-6579699bde94\") " Oct 11 04:22:49 crc kubenswrapper[4967]: I1011 04:22:49.345393 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31a11147-0dcb-4957-826d-6579699bde94-kube-api-access-hxxgw" (OuterVolumeSpecName: "kube-api-access-hxxgw") pod "31a11147-0dcb-4957-826d-6579699bde94" (UID: "31a11147-0dcb-4957-826d-6579699bde94"). InnerVolumeSpecName "kube-api-access-hxxgw". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:22:49 crc kubenswrapper[4967]: I1011 04:22:49.442020 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-hxxgw\" (UniqueName: \"kubernetes.io/projected/31a11147-0dcb-4957-826d-6579699bde94-kube-api-access-hxxgw\") on node \"crc\" DevicePath \"\"" Oct 11 04:22:49 crc kubenswrapper[4967]: I1011 04:22:49.456040 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/31a11147-0dcb-4957-826d-6579699bde94-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "31a11147-0dcb-4957-826d-6579699bde94" (UID: "31a11147-0dcb-4957-826d-6579699bde94"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:22:49 crc kubenswrapper[4967]: I1011 04:22:49.544046 4967 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/31a11147-0dcb-4957-826d-6579699bde94-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 11 04:22:49 crc kubenswrapper[4967]: I1011 04:22:49.838434 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-vt6vn_must-gather-689h5_31a11147-0dcb-4957-826d-6579699bde94/copy/0.log" Oct 11 04:22:49 crc kubenswrapper[4967]: I1011 04:22:49.839179 4967 generic.go:334] "Generic (PLEG): container finished" podID="31a11147-0dcb-4957-826d-6579699bde94" containerID="f6b72a8eac803d353e976a61f4ebfb8f5c2e6ad41329bb8a50f24479247338a7" exitCode=143 Oct 11 04:22:49 crc kubenswrapper[4967]: I1011 04:22:49.839227 4967 scope.go:117] "RemoveContainer" containerID="f6b72a8eac803d353e976a61f4ebfb8f5c2e6ad41329bb8a50f24479247338a7" Oct 11 04:22:49 crc kubenswrapper[4967]: I1011 04:22:49.839351 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-vt6vn/must-gather-689h5" Oct 11 04:22:49 crc kubenswrapper[4967]: I1011 04:22:49.875096 4967 scope.go:117] "RemoveContainer" containerID="f672d23f2106ef666680d24be5d2207af3059c78ee3b47c6c60c33f34e34e223" Oct 11 04:22:49 crc kubenswrapper[4967]: I1011 04:22:49.943955 4967 scope.go:117] "RemoveContainer" containerID="f6b72a8eac803d353e976a61f4ebfb8f5c2e6ad41329bb8a50f24479247338a7" Oct 11 04:22:49 crc kubenswrapper[4967]: E1011 04:22:49.944572 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f6b72a8eac803d353e976a61f4ebfb8f5c2e6ad41329bb8a50f24479247338a7\": container with ID starting with f6b72a8eac803d353e976a61f4ebfb8f5c2e6ad41329bb8a50f24479247338a7 not found: ID does not exist" containerID="f6b72a8eac803d353e976a61f4ebfb8f5c2e6ad41329bb8a50f24479247338a7" Oct 11 04:22:49 crc kubenswrapper[4967]: I1011 04:22:49.944618 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f6b72a8eac803d353e976a61f4ebfb8f5c2e6ad41329bb8a50f24479247338a7"} err="failed to get container status \"f6b72a8eac803d353e976a61f4ebfb8f5c2e6ad41329bb8a50f24479247338a7\": rpc error: code = NotFound desc = could not find container \"f6b72a8eac803d353e976a61f4ebfb8f5c2e6ad41329bb8a50f24479247338a7\": container with ID starting with f6b72a8eac803d353e976a61f4ebfb8f5c2e6ad41329bb8a50f24479247338a7 not found: ID does not exist" Oct 11 04:22:49 crc kubenswrapper[4967]: I1011 04:22:49.944644 4967 scope.go:117] "RemoveContainer" containerID="f672d23f2106ef666680d24be5d2207af3059c78ee3b47c6c60c33f34e34e223" Oct 11 04:22:49 crc kubenswrapper[4967]: E1011 04:22:49.945037 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f672d23f2106ef666680d24be5d2207af3059c78ee3b47c6c60c33f34e34e223\": container with ID starting with f672d23f2106ef666680d24be5d2207af3059c78ee3b47c6c60c33f34e34e223 not found: ID does not exist" containerID="f672d23f2106ef666680d24be5d2207af3059c78ee3b47c6c60c33f34e34e223" Oct 11 04:22:49 crc kubenswrapper[4967]: I1011 04:22:49.945146 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f672d23f2106ef666680d24be5d2207af3059c78ee3b47c6c60c33f34e34e223"} err="failed to get container status \"f672d23f2106ef666680d24be5d2207af3059c78ee3b47c6c60c33f34e34e223\": rpc error: code = NotFound desc = could not find container \"f672d23f2106ef666680d24be5d2207af3059c78ee3b47c6c60c33f34e34e223\": container with ID starting with f672d23f2106ef666680d24be5d2207af3059c78ee3b47c6c60c33f34e34e223 not found: ID does not exist" Oct 11 04:22:50 crc kubenswrapper[4967]: I1011 04:22:50.844124 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31a11147-0dcb-4957-826d-6579699bde94" path="/var/lib/kubelet/pods/31a11147-0dcb-4957-826d-6579699bde94/volumes" Oct 11 04:23:35 crc kubenswrapper[4967]: I1011 04:23:35.443723 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-rtxqh/must-gather-4zlrw"] Oct 11 04:23:35 crc kubenswrapper[4967]: E1011 04:23:35.447797 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31a11147-0dcb-4957-826d-6579699bde94" containerName="gather" Oct 11 04:23:35 crc kubenswrapper[4967]: I1011 04:23:35.447857 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="31a11147-0dcb-4957-826d-6579699bde94" containerName="gather" Oct 11 04:23:35 crc kubenswrapper[4967]: E1011 04:23:35.447888 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0" containerName="container-00" Oct 11 04:23:35 crc kubenswrapper[4967]: I1011 04:23:35.447903 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0" containerName="container-00" Oct 11 04:23:35 crc kubenswrapper[4967]: E1011 04:23:35.447946 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="31a11147-0dcb-4957-826d-6579699bde94" containerName="copy" Oct 11 04:23:35 crc kubenswrapper[4967]: I1011 04:23:35.447958 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="31a11147-0dcb-4957-826d-6579699bde94" containerName="copy" Oct 11 04:23:35 crc kubenswrapper[4967]: I1011 04:23:35.448571 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="31a11147-0dcb-4957-826d-6579699bde94" containerName="copy" Oct 11 04:23:35 crc kubenswrapper[4967]: I1011 04:23:35.448609 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="54bd3b93-2db2-4ea4-8bb0-ee6375bd76b0" containerName="container-00" Oct 11 04:23:35 crc kubenswrapper[4967]: I1011 04:23:35.448641 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="31a11147-0dcb-4957-826d-6579699bde94" containerName="gather" Oct 11 04:23:35 crc kubenswrapper[4967]: I1011 04:23:35.450399 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rtxqh/must-gather-4zlrw" Oct 11 04:23:35 crc kubenswrapper[4967]: I1011 04:23:35.454787 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-rtxqh"/"kube-root-ca.crt" Oct 11 04:23:35 crc kubenswrapper[4967]: I1011 04:23:35.470696 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-rtxqh"/"openshift-service-ca.crt" Oct 11 04:23:35 crc kubenswrapper[4967]: I1011 04:23:35.545788 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-rtxqh/must-gather-4zlrw"] Oct 11 04:23:35 crc kubenswrapper[4967]: I1011 04:23:35.594100 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/393350ef-d545-42cd-92d2-59ee992dd2bc-must-gather-output\") pod \"must-gather-4zlrw\" (UID: \"393350ef-d545-42cd-92d2-59ee992dd2bc\") " pod="openshift-must-gather-rtxqh/must-gather-4zlrw" Oct 11 04:23:35 crc kubenswrapper[4967]: I1011 04:23:35.594240 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q58xk\" (UniqueName: \"kubernetes.io/projected/393350ef-d545-42cd-92d2-59ee992dd2bc-kube-api-access-q58xk\") pod \"must-gather-4zlrw\" (UID: \"393350ef-d545-42cd-92d2-59ee992dd2bc\") " pod="openshift-must-gather-rtxqh/must-gather-4zlrw" Oct 11 04:23:35 crc kubenswrapper[4967]: I1011 04:23:35.696381 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q58xk\" (UniqueName: \"kubernetes.io/projected/393350ef-d545-42cd-92d2-59ee992dd2bc-kube-api-access-q58xk\") pod \"must-gather-4zlrw\" (UID: \"393350ef-d545-42cd-92d2-59ee992dd2bc\") " pod="openshift-must-gather-rtxqh/must-gather-4zlrw" Oct 11 04:23:35 crc kubenswrapper[4967]: I1011 04:23:35.696793 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/393350ef-d545-42cd-92d2-59ee992dd2bc-must-gather-output\") pod \"must-gather-4zlrw\" (UID: \"393350ef-d545-42cd-92d2-59ee992dd2bc\") " pod="openshift-must-gather-rtxqh/must-gather-4zlrw" Oct 11 04:23:35 crc kubenswrapper[4967]: I1011 04:23:35.697157 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/393350ef-d545-42cd-92d2-59ee992dd2bc-must-gather-output\") pod \"must-gather-4zlrw\" (UID: \"393350ef-d545-42cd-92d2-59ee992dd2bc\") " pod="openshift-must-gather-rtxqh/must-gather-4zlrw" Oct 11 04:23:35 crc kubenswrapper[4967]: I1011 04:23:35.712448 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q58xk\" (UniqueName: \"kubernetes.io/projected/393350ef-d545-42cd-92d2-59ee992dd2bc-kube-api-access-q58xk\") pod \"must-gather-4zlrw\" (UID: \"393350ef-d545-42cd-92d2-59ee992dd2bc\") " pod="openshift-must-gather-rtxqh/must-gather-4zlrw" Oct 11 04:23:35 crc kubenswrapper[4967]: I1011 04:23:35.775062 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rtxqh/must-gather-4zlrw" Oct 11 04:23:36 crc kubenswrapper[4967]: I1011 04:23:36.221637 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-rtxqh/must-gather-4zlrw"] Oct 11 04:23:36 crc kubenswrapper[4967]: I1011 04:23:36.342922 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rtxqh/must-gather-4zlrw" event={"ID":"393350ef-d545-42cd-92d2-59ee992dd2bc","Type":"ContainerStarted","Data":"53b96c08fd920e8e633cd292e1a87ba97459cb5ed788ab0795db468d18b40aab"} Oct 11 04:23:37 crc kubenswrapper[4967]: I1011 04:23:37.350707 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rtxqh/must-gather-4zlrw" event={"ID":"393350ef-d545-42cd-92d2-59ee992dd2bc","Type":"ContainerStarted","Data":"d53cafbbae23954dcd74b593c51989cbf17bc3f0b9eb0a7708abd835965f181a"} Oct 11 04:23:37 crc kubenswrapper[4967]: I1011 04:23:37.351651 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rtxqh/must-gather-4zlrw" event={"ID":"393350ef-d545-42cd-92d2-59ee992dd2bc","Type":"ContainerStarted","Data":"5713d3fe8a8383721a12fa412cb095b6b308a0fd981cc6327434293fca4a7482"} Oct 11 04:23:37 crc kubenswrapper[4967]: I1011 04:23:37.365682 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-rtxqh/must-gather-4zlrw" podStartSLOduration=2.365665081 podStartE2EDuration="2.365665081s" podCreationTimestamp="2025-10-11 04:23:35 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:23:37.362245521 +0000 UTC m=+1945.325454484" watchObservedRunningTime="2025-10-11 04:23:37.365665081 +0000 UTC m=+1945.328874014" Oct 11 04:23:38 crc kubenswrapper[4967]: E1011 04:23:38.904529 4967 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.50:46174->38.102.83.50:40947: write tcp 38.102.83.50:46174->38.102.83.50:40947: write: broken pipe Oct 11 04:23:40 crc kubenswrapper[4967]: I1011 04:23:40.686266 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-rtxqh/crc-debug-4684w"] Oct 11 04:23:40 crc kubenswrapper[4967]: I1011 04:23:40.688285 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rtxqh/crc-debug-4684w" Oct 11 04:23:40 crc kubenswrapper[4967]: I1011 04:23:40.690349 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-rtxqh"/"default-dockercfg-zpv2n" Oct 11 04:23:40 crc kubenswrapper[4967]: I1011 04:23:40.797094 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xblxn\" (UniqueName: \"kubernetes.io/projected/9eaa3442-aa1f-4a79-a9b7-75dd50f3829a-kube-api-access-xblxn\") pod \"crc-debug-4684w\" (UID: \"9eaa3442-aa1f-4a79-a9b7-75dd50f3829a\") " pod="openshift-must-gather-rtxqh/crc-debug-4684w" Oct 11 04:23:40 crc kubenswrapper[4967]: I1011 04:23:40.797258 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9eaa3442-aa1f-4a79-a9b7-75dd50f3829a-host\") pod \"crc-debug-4684w\" (UID: \"9eaa3442-aa1f-4a79-a9b7-75dd50f3829a\") " pod="openshift-must-gather-rtxqh/crc-debug-4684w" Oct 11 04:23:40 crc kubenswrapper[4967]: I1011 04:23:40.899461 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9eaa3442-aa1f-4a79-a9b7-75dd50f3829a-host\") pod \"crc-debug-4684w\" (UID: \"9eaa3442-aa1f-4a79-a9b7-75dd50f3829a\") " pod="openshift-must-gather-rtxqh/crc-debug-4684w" Oct 11 04:23:40 crc kubenswrapper[4967]: I1011 04:23:40.899574 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xblxn\" (UniqueName: \"kubernetes.io/projected/9eaa3442-aa1f-4a79-a9b7-75dd50f3829a-kube-api-access-xblxn\") pod \"crc-debug-4684w\" (UID: \"9eaa3442-aa1f-4a79-a9b7-75dd50f3829a\") " pod="openshift-must-gather-rtxqh/crc-debug-4684w" Oct 11 04:23:40 crc kubenswrapper[4967]: I1011 04:23:40.899662 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9eaa3442-aa1f-4a79-a9b7-75dd50f3829a-host\") pod \"crc-debug-4684w\" (UID: \"9eaa3442-aa1f-4a79-a9b7-75dd50f3829a\") " pod="openshift-must-gather-rtxqh/crc-debug-4684w" Oct 11 04:23:40 crc kubenswrapper[4967]: I1011 04:23:40.917930 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xblxn\" (UniqueName: \"kubernetes.io/projected/9eaa3442-aa1f-4a79-a9b7-75dd50f3829a-kube-api-access-xblxn\") pod \"crc-debug-4684w\" (UID: \"9eaa3442-aa1f-4a79-a9b7-75dd50f3829a\") " pod="openshift-must-gather-rtxqh/crc-debug-4684w" Oct 11 04:23:41 crc kubenswrapper[4967]: I1011 04:23:41.014352 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rtxqh/crc-debug-4684w" Oct 11 04:23:41 crc kubenswrapper[4967]: W1011 04:23:41.047399 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod9eaa3442_aa1f_4a79_a9b7_75dd50f3829a.slice/crio-015a1c1bde55b8b99f50766c4d783816213d40835f1a99c76705c6326166db1e WatchSource:0}: Error finding container 015a1c1bde55b8b99f50766c4d783816213d40835f1a99c76705c6326166db1e: Status 404 returned error can't find the container with id 015a1c1bde55b8b99f50766c4d783816213d40835f1a99c76705c6326166db1e Oct 11 04:23:41 crc kubenswrapper[4967]: I1011 04:23:41.390954 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rtxqh/crc-debug-4684w" event={"ID":"9eaa3442-aa1f-4a79-a9b7-75dd50f3829a","Type":"ContainerStarted","Data":"d01ccfc202d6e2c3e55352f3ed8bf385f61af024595e9dee56cb5a918d76cd6c"} Oct 11 04:23:41 crc kubenswrapper[4967]: I1011 04:23:41.391329 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rtxqh/crc-debug-4684w" event={"ID":"9eaa3442-aa1f-4a79-a9b7-75dd50f3829a","Type":"ContainerStarted","Data":"015a1c1bde55b8b99f50766c4d783816213d40835f1a99c76705c6326166db1e"} Oct 11 04:23:41 crc kubenswrapper[4967]: I1011 04:23:41.408875 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-rtxqh/crc-debug-4684w" podStartSLOduration=1.408859216 podStartE2EDuration="1.408859216s" podCreationTimestamp="2025-10-11 04:23:40 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:23:41.403922252 +0000 UTC m=+1949.367131185" watchObservedRunningTime="2025-10-11 04:23:41.408859216 +0000 UTC m=+1949.372068149" Oct 11 04:23:44 crc kubenswrapper[4967]: I1011 04:23:44.088735 4967 patch_prober.go:28] interesting pod/machine-config-daemon-sgrv2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:23:44 crc kubenswrapper[4967]: I1011 04:23:44.089177 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:24:13 crc kubenswrapper[4967]: I1011 04:24:13.690820 4967 generic.go:334] "Generic (PLEG): container finished" podID="9eaa3442-aa1f-4a79-a9b7-75dd50f3829a" containerID="d01ccfc202d6e2c3e55352f3ed8bf385f61af024595e9dee56cb5a918d76cd6c" exitCode=0 Oct 11 04:24:13 crc kubenswrapper[4967]: I1011 04:24:13.690905 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rtxqh/crc-debug-4684w" event={"ID":"9eaa3442-aa1f-4a79-a9b7-75dd50f3829a","Type":"ContainerDied","Data":"d01ccfc202d6e2c3e55352f3ed8bf385f61af024595e9dee56cb5a918d76cd6c"} Oct 11 04:24:14 crc kubenswrapper[4967]: I1011 04:24:14.089177 4967 patch_prober.go:28] interesting pod/machine-config-daemon-sgrv2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:24:14 crc kubenswrapper[4967]: I1011 04:24:14.089233 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:24:14 crc kubenswrapper[4967]: I1011 04:24:14.827958 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rtxqh/crc-debug-4684w" Oct 11 04:24:14 crc kubenswrapper[4967]: I1011 04:24:14.870583 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-rtxqh/crc-debug-4684w"] Oct 11 04:24:14 crc kubenswrapper[4967]: I1011 04:24:14.879626 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-rtxqh/crc-debug-4684w"] Oct 11 04:24:14 crc kubenswrapper[4967]: I1011 04:24:14.929295 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xblxn\" (UniqueName: \"kubernetes.io/projected/9eaa3442-aa1f-4a79-a9b7-75dd50f3829a-kube-api-access-xblxn\") pod \"9eaa3442-aa1f-4a79-a9b7-75dd50f3829a\" (UID: \"9eaa3442-aa1f-4a79-a9b7-75dd50f3829a\") " Oct 11 04:24:14 crc kubenswrapper[4967]: I1011 04:24:14.929591 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9eaa3442-aa1f-4a79-a9b7-75dd50f3829a-host\") pod \"9eaa3442-aa1f-4a79-a9b7-75dd50f3829a\" (UID: \"9eaa3442-aa1f-4a79-a9b7-75dd50f3829a\") " Oct 11 04:24:14 crc kubenswrapper[4967]: I1011 04:24:14.929666 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/9eaa3442-aa1f-4a79-a9b7-75dd50f3829a-host" (OuterVolumeSpecName: "host") pod "9eaa3442-aa1f-4a79-a9b7-75dd50f3829a" (UID: "9eaa3442-aa1f-4a79-a9b7-75dd50f3829a"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 04:24:14 crc kubenswrapper[4967]: I1011 04:24:14.930785 4967 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/9eaa3442-aa1f-4a79-a9b7-75dd50f3829a-host\") on node \"crc\" DevicePath \"\"" Oct 11 04:24:14 crc kubenswrapper[4967]: I1011 04:24:14.936100 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9eaa3442-aa1f-4a79-a9b7-75dd50f3829a-kube-api-access-xblxn" (OuterVolumeSpecName: "kube-api-access-xblxn") pod "9eaa3442-aa1f-4a79-a9b7-75dd50f3829a" (UID: "9eaa3442-aa1f-4a79-a9b7-75dd50f3829a"). InnerVolumeSpecName "kube-api-access-xblxn". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:24:15 crc kubenswrapper[4967]: I1011 04:24:15.032727 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xblxn\" (UniqueName: \"kubernetes.io/projected/9eaa3442-aa1f-4a79-a9b7-75dd50f3829a-kube-api-access-xblxn\") on node \"crc\" DevicePath \"\"" Oct 11 04:24:15 crc kubenswrapper[4967]: I1011 04:24:15.713463 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="015a1c1bde55b8b99f50766c4d783816213d40835f1a99c76705c6326166db1e" Oct 11 04:24:15 crc kubenswrapper[4967]: I1011 04:24:15.713529 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rtxqh/crc-debug-4684w" Oct 11 04:24:16 crc kubenswrapper[4967]: I1011 04:24:16.097528 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-rtxqh/crc-debug-8fzrf"] Oct 11 04:24:16 crc kubenswrapper[4967]: E1011 04:24:16.098104 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9eaa3442-aa1f-4a79-a9b7-75dd50f3829a" containerName="container-00" Oct 11 04:24:16 crc kubenswrapper[4967]: I1011 04:24:16.098121 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="9eaa3442-aa1f-4a79-a9b7-75dd50f3829a" containerName="container-00" Oct 11 04:24:16 crc kubenswrapper[4967]: I1011 04:24:16.098403 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="9eaa3442-aa1f-4a79-a9b7-75dd50f3829a" containerName="container-00" Oct 11 04:24:16 crc kubenswrapper[4967]: I1011 04:24:16.099311 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rtxqh/crc-debug-8fzrf" Oct 11 04:24:16 crc kubenswrapper[4967]: I1011 04:24:16.103485 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-must-gather-rtxqh"/"default-dockercfg-zpv2n" Oct 11 04:24:16 crc kubenswrapper[4967]: I1011 04:24:16.158754 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/137cc419-0c56-4783-9864-dff1f4e72a21-host\") pod \"crc-debug-8fzrf\" (UID: \"137cc419-0c56-4783-9864-dff1f4e72a21\") " pod="openshift-must-gather-rtxqh/crc-debug-8fzrf" Oct 11 04:24:16 crc kubenswrapper[4967]: I1011 04:24:16.158812 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-65gh7\" (UniqueName: \"kubernetes.io/projected/137cc419-0c56-4783-9864-dff1f4e72a21-kube-api-access-65gh7\") pod \"crc-debug-8fzrf\" (UID: \"137cc419-0c56-4783-9864-dff1f4e72a21\") " pod="openshift-must-gather-rtxqh/crc-debug-8fzrf" Oct 11 04:24:16 crc kubenswrapper[4967]: I1011 04:24:16.260633 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/137cc419-0c56-4783-9864-dff1f4e72a21-host\") pod \"crc-debug-8fzrf\" (UID: \"137cc419-0c56-4783-9864-dff1f4e72a21\") " pod="openshift-must-gather-rtxqh/crc-debug-8fzrf" Oct 11 04:24:16 crc kubenswrapper[4967]: I1011 04:24:16.261036 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-65gh7\" (UniqueName: \"kubernetes.io/projected/137cc419-0c56-4783-9864-dff1f4e72a21-kube-api-access-65gh7\") pod \"crc-debug-8fzrf\" (UID: \"137cc419-0c56-4783-9864-dff1f4e72a21\") " pod="openshift-must-gather-rtxqh/crc-debug-8fzrf" Oct 11 04:24:16 crc kubenswrapper[4967]: I1011 04:24:16.261797 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/137cc419-0c56-4783-9864-dff1f4e72a21-host\") pod \"crc-debug-8fzrf\" (UID: \"137cc419-0c56-4783-9864-dff1f4e72a21\") " pod="openshift-must-gather-rtxqh/crc-debug-8fzrf" Oct 11 04:24:16 crc kubenswrapper[4967]: I1011 04:24:16.280735 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-65gh7\" (UniqueName: \"kubernetes.io/projected/137cc419-0c56-4783-9864-dff1f4e72a21-kube-api-access-65gh7\") pod \"crc-debug-8fzrf\" (UID: \"137cc419-0c56-4783-9864-dff1f4e72a21\") " pod="openshift-must-gather-rtxqh/crc-debug-8fzrf" Oct 11 04:24:16 crc kubenswrapper[4967]: I1011 04:24:16.419983 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rtxqh/crc-debug-8fzrf" Oct 11 04:24:16 crc kubenswrapper[4967]: W1011 04:24:16.449390 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod137cc419_0c56_4783_9864_dff1f4e72a21.slice/crio-c8cfdf5de31b8827f3305715c2cc9000d63380b5bea8540857750b28cf65c2a7 WatchSource:0}: Error finding container c8cfdf5de31b8827f3305715c2cc9000d63380b5bea8540857750b28cf65c2a7: Status 404 returned error can't find the container with id c8cfdf5de31b8827f3305715c2cc9000d63380b5bea8540857750b28cf65c2a7 Oct 11 04:24:16 crc kubenswrapper[4967]: I1011 04:24:16.723608 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rtxqh/crc-debug-8fzrf" event={"ID":"137cc419-0c56-4783-9864-dff1f4e72a21","Type":"ContainerStarted","Data":"ec566115d44487ea28c71c30b91b2af67daa7ed30fbb0b1042dadcc8f3ba28d4"} Oct 11 04:24:16 crc kubenswrapper[4967]: I1011 04:24:16.723947 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rtxqh/crc-debug-8fzrf" event={"ID":"137cc419-0c56-4783-9864-dff1f4e72a21","Type":"ContainerStarted","Data":"c8cfdf5de31b8827f3305715c2cc9000d63380b5bea8540857750b28cf65c2a7"} Oct 11 04:24:16 crc kubenswrapper[4967]: I1011 04:24:16.826405 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9eaa3442-aa1f-4a79-a9b7-75dd50f3829a" path="/var/lib/kubelet/pods/9eaa3442-aa1f-4a79-a9b7-75dd50f3829a/volumes" Oct 11 04:24:17 crc kubenswrapper[4967]: I1011 04:24:17.732975 4967 generic.go:334] "Generic (PLEG): container finished" podID="137cc419-0c56-4783-9864-dff1f4e72a21" containerID="ec566115d44487ea28c71c30b91b2af67daa7ed30fbb0b1042dadcc8f3ba28d4" exitCode=0 Oct 11 04:24:17 crc kubenswrapper[4967]: I1011 04:24:17.733041 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rtxqh/crc-debug-8fzrf" event={"ID":"137cc419-0c56-4783-9864-dff1f4e72a21","Type":"ContainerDied","Data":"ec566115d44487ea28c71c30b91b2af67daa7ed30fbb0b1042dadcc8f3ba28d4"} Oct 11 04:24:18 crc kubenswrapper[4967]: I1011 04:24:18.248266 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-rtxqh/crc-debug-8fzrf"] Oct 11 04:24:18 crc kubenswrapper[4967]: I1011 04:24:18.256851 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-rtxqh/crc-debug-8fzrf"] Oct 11 04:24:18 crc kubenswrapper[4967]: I1011 04:24:18.846473 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rtxqh/crc-debug-8fzrf" Oct 11 04:24:19 crc kubenswrapper[4967]: I1011 04:24:19.013438 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-65gh7\" (UniqueName: \"kubernetes.io/projected/137cc419-0c56-4783-9864-dff1f4e72a21-kube-api-access-65gh7\") pod \"137cc419-0c56-4783-9864-dff1f4e72a21\" (UID: \"137cc419-0c56-4783-9864-dff1f4e72a21\") " Oct 11 04:24:19 crc kubenswrapper[4967]: I1011 04:24:19.013511 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/137cc419-0c56-4783-9864-dff1f4e72a21-host\") pod \"137cc419-0c56-4783-9864-dff1f4e72a21\" (UID: \"137cc419-0c56-4783-9864-dff1f4e72a21\") " Oct 11 04:24:19 crc kubenswrapper[4967]: I1011 04:24:19.013724 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/137cc419-0c56-4783-9864-dff1f4e72a21-host" (OuterVolumeSpecName: "host") pod "137cc419-0c56-4783-9864-dff1f4e72a21" (UID: "137cc419-0c56-4783-9864-dff1f4e72a21"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 04:24:19 crc kubenswrapper[4967]: I1011 04:24:19.014330 4967 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/137cc419-0c56-4783-9864-dff1f4e72a21-host\") on node \"crc\" DevicePath \"\"" Oct 11 04:24:19 crc kubenswrapper[4967]: I1011 04:24:19.023281 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/137cc419-0c56-4783-9864-dff1f4e72a21-kube-api-access-65gh7" (OuterVolumeSpecName: "kube-api-access-65gh7") pod "137cc419-0c56-4783-9864-dff1f4e72a21" (UID: "137cc419-0c56-4783-9864-dff1f4e72a21"). InnerVolumeSpecName "kube-api-access-65gh7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:24:19 crc kubenswrapper[4967]: I1011 04:24:19.116448 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-65gh7\" (UniqueName: \"kubernetes.io/projected/137cc419-0c56-4783-9864-dff1f4e72a21-kube-api-access-65gh7\") on node \"crc\" DevicePath \"\"" Oct 11 04:24:19 crc kubenswrapper[4967]: I1011 04:24:19.470848 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-rtxqh/crc-debug-kkkbn"] Oct 11 04:24:19 crc kubenswrapper[4967]: E1011 04:24:19.471483 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="137cc419-0c56-4783-9864-dff1f4e72a21" containerName="container-00" Oct 11 04:24:19 crc kubenswrapper[4967]: I1011 04:24:19.471499 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="137cc419-0c56-4783-9864-dff1f4e72a21" containerName="container-00" Oct 11 04:24:19 crc kubenswrapper[4967]: I1011 04:24:19.471700 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="137cc419-0c56-4783-9864-dff1f4e72a21" containerName="container-00" Oct 11 04:24:19 crc kubenswrapper[4967]: I1011 04:24:19.472323 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rtxqh/crc-debug-kkkbn" Oct 11 04:24:19 crc kubenswrapper[4967]: I1011 04:24:19.626561 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-txgbt\" (UniqueName: \"kubernetes.io/projected/5b8a9b80-ec80-4d85-bada-f0c1420e0f3f-kube-api-access-txgbt\") pod \"crc-debug-kkkbn\" (UID: \"5b8a9b80-ec80-4d85-bada-f0c1420e0f3f\") " pod="openshift-must-gather-rtxqh/crc-debug-kkkbn" Oct 11 04:24:19 crc kubenswrapper[4967]: I1011 04:24:19.626719 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5b8a9b80-ec80-4d85-bada-f0c1420e0f3f-host\") pod \"crc-debug-kkkbn\" (UID: \"5b8a9b80-ec80-4d85-bada-f0c1420e0f3f\") " pod="openshift-must-gather-rtxqh/crc-debug-kkkbn" Oct 11 04:24:19 crc kubenswrapper[4967]: I1011 04:24:19.728569 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-txgbt\" (UniqueName: \"kubernetes.io/projected/5b8a9b80-ec80-4d85-bada-f0c1420e0f3f-kube-api-access-txgbt\") pod \"crc-debug-kkkbn\" (UID: \"5b8a9b80-ec80-4d85-bada-f0c1420e0f3f\") " pod="openshift-must-gather-rtxqh/crc-debug-kkkbn" Oct 11 04:24:19 crc kubenswrapper[4967]: I1011 04:24:19.728689 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5b8a9b80-ec80-4d85-bada-f0c1420e0f3f-host\") pod \"crc-debug-kkkbn\" (UID: \"5b8a9b80-ec80-4d85-bada-f0c1420e0f3f\") " pod="openshift-must-gather-rtxqh/crc-debug-kkkbn" Oct 11 04:24:19 crc kubenswrapper[4967]: I1011 04:24:19.728801 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5b8a9b80-ec80-4d85-bada-f0c1420e0f3f-host\") pod \"crc-debug-kkkbn\" (UID: \"5b8a9b80-ec80-4d85-bada-f0c1420e0f3f\") " pod="openshift-must-gather-rtxqh/crc-debug-kkkbn" Oct 11 04:24:19 crc kubenswrapper[4967]: I1011 04:24:19.748146 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-txgbt\" (UniqueName: \"kubernetes.io/projected/5b8a9b80-ec80-4d85-bada-f0c1420e0f3f-kube-api-access-txgbt\") pod \"crc-debug-kkkbn\" (UID: \"5b8a9b80-ec80-4d85-bada-f0c1420e0f3f\") " pod="openshift-must-gather-rtxqh/crc-debug-kkkbn" Oct 11 04:24:19 crc kubenswrapper[4967]: I1011 04:24:19.752330 4967 scope.go:117] "RemoveContainer" containerID="ec566115d44487ea28c71c30b91b2af67daa7ed30fbb0b1042dadcc8f3ba28d4" Oct 11 04:24:19 crc kubenswrapper[4967]: I1011 04:24:19.752504 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rtxqh/crc-debug-8fzrf" Oct 11 04:24:19 crc kubenswrapper[4967]: I1011 04:24:19.791020 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rtxqh/crc-debug-kkkbn" Oct 11 04:24:20 crc kubenswrapper[4967]: I1011 04:24:20.767683 4967 generic.go:334] "Generic (PLEG): container finished" podID="5b8a9b80-ec80-4d85-bada-f0c1420e0f3f" containerID="6fcdda454c0c7f453665551fd9686dbd65bfe089b711fdd55bc4b7b27ac472f7" exitCode=0 Oct 11 04:24:20 crc kubenswrapper[4967]: I1011 04:24:20.768012 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rtxqh/crc-debug-kkkbn" event={"ID":"5b8a9b80-ec80-4d85-bada-f0c1420e0f3f","Type":"ContainerDied","Data":"6fcdda454c0c7f453665551fd9686dbd65bfe089b711fdd55bc4b7b27ac472f7"} Oct 11 04:24:20 crc kubenswrapper[4967]: I1011 04:24:20.768042 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rtxqh/crc-debug-kkkbn" event={"ID":"5b8a9b80-ec80-4d85-bada-f0c1420e0f3f","Type":"ContainerStarted","Data":"85cd39f0eb5e388fba3672e8688c2090c22f439ccaa6c1289314238b87488394"} Oct 11 04:24:20 crc kubenswrapper[4967]: I1011 04:24:20.826632 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="137cc419-0c56-4783-9864-dff1f4e72a21" path="/var/lib/kubelet/pods/137cc419-0c56-4783-9864-dff1f4e72a21/volumes" Oct 11 04:24:20 crc kubenswrapper[4967]: I1011 04:24:20.827187 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-rtxqh/crc-debug-kkkbn"] Oct 11 04:24:20 crc kubenswrapper[4967]: I1011 04:24:20.827220 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-rtxqh/crc-debug-kkkbn"] Oct 11 04:24:21 crc kubenswrapper[4967]: I1011 04:24:21.877492 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rtxqh/crc-debug-kkkbn" Oct 11 04:24:22 crc kubenswrapper[4967]: I1011 04:24:22.073093 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-txgbt\" (UniqueName: \"kubernetes.io/projected/5b8a9b80-ec80-4d85-bada-f0c1420e0f3f-kube-api-access-txgbt\") pod \"5b8a9b80-ec80-4d85-bada-f0c1420e0f3f\" (UID: \"5b8a9b80-ec80-4d85-bada-f0c1420e0f3f\") " Oct 11 04:24:22 crc kubenswrapper[4967]: I1011 04:24:22.073393 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5b8a9b80-ec80-4d85-bada-f0c1420e0f3f-host\") pod \"5b8a9b80-ec80-4d85-bada-f0c1420e0f3f\" (UID: \"5b8a9b80-ec80-4d85-bada-f0c1420e0f3f\") " Oct 11 04:24:22 crc kubenswrapper[4967]: I1011 04:24:22.073989 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/5b8a9b80-ec80-4d85-bada-f0c1420e0f3f-host" (OuterVolumeSpecName: "host") pod "5b8a9b80-ec80-4d85-bada-f0c1420e0f3f" (UID: "5b8a9b80-ec80-4d85-bada-f0c1420e0f3f"). InnerVolumeSpecName "host". PluginName "kubernetes.io/host-path", VolumeGidValue "" Oct 11 04:24:22 crc kubenswrapper[4967]: I1011 04:24:22.078986 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b8a9b80-ec80-4d85-bada-f0c1420e0f3f-kube-api-access-txgbt" (OuterVolumeSpecName: "kube-api-access-txgbt") pod "5b8a9b80-ec80-4d85-bada-f0c1420e0f3f" (UID: "5b8a9b80-ec80-4d85-bada-f0c1420e0f3f"). InnerVolumeSpecName "kube-api-access-txgbt". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:24:22 crc kubenswrapper[4967]: I1011 04:24:22.175635 4967 reconciler_common.go:293] "Volume detached for volume \"host\" (UniqueName: \"kubernetes.io/host-path/5b8a9b80-ec80-4d85-bada-f0c1420e0f3f-host\") on node \"crc\" DevicePath \"\"" Oct 11 04:24:22 crc kubenswrapper[4967]: I1011 04:24:22.175979 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-txgbt\" (UniqueName: \"kubernetes.io/projected/5b8a9b80-ec80-4d85-bada-f0c1420e0f3f-kube-api-access-txgbt\") on node \"crc\" DevicePath \"\"" Oct 11 04:24:22 crc kubenswrapper[4967]: I1011 04:24:22.787188 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="85cd39f0eb5e388fba3672e8688c2090c22f439ccaa6c1289314238b87488394" Oct 11 04:24:22 crc kubenswrapper[4967]: I1011 04:24:22.787261 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rtxqh/crc-debug-kkkbn" Oct 11 04:24:22 crc kubenswrapper[4967]: I1011 04:24:22.829776 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b8a9b80-ec80-4d85-bada-f0c1420e0f3f" path="/var/lib/kubelet/pods/5b8a9b80-ec80-4d85-bada-f0c1420e0f3f/volumes" Oct 11 04:24:34 crc kubenswrapper[4967]: I1011 04:24:34.779515 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-84674df998-htlks_c72a2b43-59ac-450c-87d3-f16c9ee4afa0/barbican-api/0.log" Oct 11 04:24:34 crc kubenswrapper[4967]: I1011 04:24:34.924403 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-api-84674df998-htlks_c72a2b43-59ac-450c-87d3-f16c9ee4afa0/barbican-api-log/0.log" Oct 11 04:24:34 crc kubenswrapper[4967]: I1011 04:24:34.969616 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7769b9cccd-zkh5v_38e5cf89-c793-4207-b06e-d117da3dd152/barbican-keystone-listener/0.log" Oct 11 04:24:35 crc kubenswrapper[4967]: I1011 04:24:35.114929 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-keystone-listener-7769b9cccd-zkh5v_38e5cf89-c793-4207-b06e-d117da3dd152/barbican-keystone-listener-log/0.log" Oct 11 04:24:35 crc kubenswrapper[4967]: I1011 04:24:35.148404 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6c87f65cff-p6j2x_fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35/barbican-worker/0.log" Oct 11 04:24:35 crc kubenswrapper[4967]: I1011 04:24:35.227968 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_barbican-worker-6c87f65cff-p6j2x_fd1c5b3c-bcb3-46cb-b4d5-3fe92fe5ee35/barbican-worker-log/0.log" Oct 11 04:24:35 crc kubenswrapper[4967]: I1011 04:24:35.356038 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c47c6348-b68f-410e-93d7-c10eed165a34/ceilometer-notification-agent/0.log" Oct 11 04:24:35 crc kubenswrapper[4967]: I1011 04:24:35.361563 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c47c6348-b68f-410e-93d7-c10eed165a34/ceilometer-central-agent/0.log" Oct 11 04:24:35 crc kubenswrapper[4967]: I1011 04:24:35.423256 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c47c6348-b68f-410e-93d7-c10eed165a34/proxy-httpd/0.log" Oct 11 04:24:35 crc kubenswrapper[4967]: I1011 04:24:35.494559 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ceilometer-0_c47c6348-b68f-410e-93d7-c10eed165a34/sg-core/0.log" Oct 11 04:24:35 crc kubenswrapper[4967]: I1011 04:24:35.618779 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_c900849f-54f8-44ed-b759-ae6fbbe441da/cinder-api-log/0.log" Oct 11 04:24:35 crc kubenswrapper[4967]: I1011 04:24:35.632184 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-api-0_c900849f-54f8-44ed-b759-ae6fbbe441da/cinder-api/0.log" Oct 11 04:24:35 crc kubenswrapper[4967]: I1011 04:24:35.805886 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_a824a183-d65e-4b45-8ecf-72bd3ce9063b/probe/0.log" Oct 11 04:24:35 crc kubenswrapper[4967]: I1011 04:24:35.851691 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_cinder-scheduler-0_a824a183-d65e-4b45-8ecf-72bd3ce9063b/cinder-scheduler/0.log" Oct 11 04:24:35 crc kubenswrapper[4967]: I1011 04:24:35.943678 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cd5cbd7b9-87xhd_ef02fdf0-4696-42b9-9095-b000a70fe334/init/0.log" Oct 11 04:24:36 crc kubenswrapper[4967]: I1011 04:24:36.097732 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cd5cbd7b9-87xhd_ef02fdf0-4696-42b9-9095-b000a70fe334/init/0.log" Oct 11 04:24:36 crc kubenswrapper[4967]: I1011 04:24:36.112668 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_dnsmasq-dns-cd5cbd7b9-87xhd_ef02fdf0-4696-42b9-9095-b000a70fe334/dnsmasq-dns/0.log" Oct 11 04:24:36 crc kubenswrapper[4967]: I1011 04:24:36.154029 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_662b8e47-db97-4003-892a-48c25cba4589/glance-httpd/0.log" Oct 11 04:24:36 crc kubenswrapper[4967]: I1011 04:24:36.316385 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-external-api-0_662b8e47-db97-4003-892a-48c25cba4589/glance-log/0.log" Oct 11 04:24:36 crc kubenswrapper[4967]: I1011 04:24:36.372035 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_53708cdd-1297-4302-9d8a-e00ba57c01b3/glance-log/0.log" Oct 11 04:24:36 crc kubenswrapper[4967]: I1011 04:24:36.397282 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_glance-default-internal-api-0_53708cdd-1297-4302-9d8a-e00ba57c01b3/glance-httpd/0.log" Oct 11 04:24:36 crc kubenswrapper[4967]: I1011 04:24:36.553832 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-85b58c7f48-9dgpg_f8ec340a-d11c-4b3b-9d2c-691df43b8892/init/0.log" Oct 11 04:24:36 crc kubenswrapper[4967]: I1011 04:24:36.695470 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-85b58c7f48-9dgpg_f8ec340a-d11c-4b3b-9d2c-691df43b8892/init/0.log" Oct 11 04:24:36 crc kubenswrapper[4967]: I1011 04:24:36.700097 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-85b58c7f48-9dgpg_f8ec340a-d11c-4b3b-9d2c-691df43b8892/ironic-api-log/0.log" Oct 11 04:24:36 crc kubenswrapper[4967]: I1011 04:24:36.779714 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-85b58c7f48-9dgpg_f8ec340a-d11c-4b3b-9d2c-691df43b8892/ironic-api/0.log" Oct 11 04:24:36 crc kubenswrapper[4967]: I1011 04:24:36.914209 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_64c75ea6-aed6-4bf6-8eef-39e76384e14f/init/0.log" Oct 11 04:24:37 crc kubenswrapper[4967]: I1011 04:24:37.018264 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_64c75ea6-aed6-4bf6-8eef-39e76384e14f/init/0.log" Oct 11 04:24:37 crc kubenswrapper[4967]: I1011 04:24:37.051975 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_64c75ea6-aed6-4bf6-8eef-39e76384e14f/ironic-python-agent-init/0.log" Oct 11 04:24:37 crc kubenswrapper[4967]: I1011 04:24:37.144008 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_64c75ea6-aed6-4bf6-8eef-39e76384e14f/ironic-python-agent-init/0.log" Oct 11 04:24:37 crc kubenswrapper[4967]: I1011 04:24:37.322286 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_64c75ea6-aed6-4bf6-8eef-39e76384e14f/init/0.log" Oct 11 04:24:37 crc kubenswrapper[4967]: I1011 04:24:37.331343 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_64c75ea6-aed6-4bf6-8eef-39e76384e14f/ironic-python-agent-init/0.log" Oct 11 04:24:37 crc kubenswrapper[4967]: I1011 04:24:37.749801 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_64c75ea6-aed6-4bf6-8eef-39e76384e14f/init/0.log" Oct 11 04:24:37 crc kubenswrapper[4967]: I1011 04:24:37.960522 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_64c75ea6-aed6-4bf6-8eef-39e76384e14f/ironic-python-agent-init/0.log" Oct 11 04:24:38 crc kubenswrapper[4967]: I1011 04:24:38.138554 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_64c75ea6-aed6-4bf6-8eef-39e76384e14f/pxe-init/0.log" Oct 11 04:24:38 crc kubenswrapper[4967]: I1011 04:24:38.147772 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_64c75ea6-aed6-4bf6-8eef-39e76384e14f/httpboot/0.log" Oct 11 04:24:38 crc kubenswrapper[4967]: I1011 04:24:38.442269 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_64c75ea6-aed6-4bf6-8eef-39e76384e14f/ramdisk-logs/0.log" Oct 11 04:24:38 crc kubenswrapper[4967]: I1011 04:24:38.519755 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_64c75ea6-aed6-4bf6-8eef-39e76384e14f/ironic-conductor/0.log" Oct 11 04:24:38 crc kubenswrapper[4967]: I1011 04:24:38.723881 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-db-sync-sdg2s_1d249b36-580e-424d-8f36-e366c932411c/init/0.log" Oct 11 04:24:38 crc kubenswrapper[4967]: I1011 04:24:38.767821 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_64c75ea6-aed6-4bf6-8eef-39e76384e14f/pxe-init/0.log" Oct 11 04:24:38 crc kubenswrapper[4967]: I1011 04:24:38.863593 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_64c75ea6-aed6-4bf6-8eef-39e76384e14f/pxe-init/0.log" Oct 11 04:24:38 crc kubenswrapper[4967]: I1011 04:24:38.878502 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-db-sync-sdg2s_1d249b36-580e-424d-8f36-e366c932411c/init/0.log" Oct 11 04:24:38 crc kubenswrapper[4967]: I1011 04:24:38.960973 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-db-sync-sdg2s_1d249b36-580e-424d-8f36-e366c932411c/ironic-db-sync/0.log" Oct 11 04:24:39 crc kubenswrapper[4967]: I1011 04:24:39.045695 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_c619427d-62f7-463c-8262-d3b9fdd6ade0/ironic-python-agent-init/0.log" Oct 11 04:24:39 crc kubenswrapper[4967]: I1011 04:24:39.129740 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-conductor-0_64c75ea6-aed6-4bf6-8eef-39e76384e14f/pxe-init/0.log" Oct 11 04:24:39 crc kubenswrapper[4967]: I1011 04:24:39.292479 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_c619427d-62f7-463c-8262-d3b9fdd6ade0/inspector-pxe-init/0.log" Oct 11 04:24:39 crc kubenswrapper[4967]: I1011 04:24:39.298245 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_c619427d-62f7-463c-8262-d3b9fdd6ade0/ironic-python-agent-init/0.log" Oct 11 04:24:39 crc kubenswrapper[4967]: I1011 04:24:39.306002 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_c619427d-62f7-463c-8262-d3b9fdd6ade0/inspector-pxe-init/0.log" Oct 11 04:24:39 crc kubenswrapper[4967]: I1011 04:24:39.454746 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_c619427d-62f7-463c-8262-d3b9fdd6ade0/ironic-python-agent-init/0.log" Oct 11 04:24:39 crc kubenswrapper[4967]: I1011 04:24:39.491117 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_c619427d-62f7-463c-8262-d3b9fdd6ade0/inspector-httpboot/0.log" Oct 11 04:24:39 crc kubenswrapper[4967]: I1011 04:24:39.494391 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_c619427d-62f7-463c-8262-d3b9fdd6ade0/inspector-pxe-init/0.log" Oct 11 04:24:39 crc kubenswrapper[4967]: I1011 04:24:39.519391 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_c619427d-62f7-463c-8262-d3b9fdd6ade0/ironic-inspector/0.log" Oct 11 04:24:39 crc kubenswrapper[4967]: I1011 04:24:39.547295 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_c619427d-62f7-463c-8262-d3b9fdd6ade0/ironic-inspector-httpd/0.log" Oct 11 04:24:39 crc kubenswrapper[4967]: I1011 04:24:39.669539 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-db-sync-m5k2b_1d9749c8-f6a4-497f-b6dd-794865243379/ironic-inspector-db-sync/0.log" Oct 11 04:24:39 crc kubenswrapper[4967]: I1011 04:24:39.683513 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-inspector-0_c619427d-62f7-463c-8262-d3b9fdd6ade0/ramdisk-logs/0.log" Oct 11 04:24:39 crc kubenswrapper[4967]: I1011 04:24:39.741971 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-neutron-agent-657f6548d9-9fb8w_4390a950-7185-4fc0-b152-b5927940032b/ironic-neutron-agent/3.log" Oct 11 04:24:39 crc kubenswrapper[4967]: I1011 04:24:39.855766 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ironic-neutron-agent-657f6548d9-9fb8w_4390a950-7185-4fc0-b152-b5927940032b/ironic-neutron-agent/2.log" Oct 11 04:24:39 crc kubenswrapper[4967]: I1011 04:24:39.939405 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_kube-state-metrics-0_f5917a28-16f9-4b7b-8b12-04c44c5333cf/kube-state-metrics/0.log" Oct 11 04:24:40 crc kubenswrapper[4967]: I1011 04:24:40.128208 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_keystone-76b6798d84-nx4fq_eb7b8d58-ee2e-4203-a1a9-44b2af5599cb/keystone-api/0.log" Oct 11 04:24:40 crc kubenswrapper[4967]: I1011 04:24:40.349564 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5f96fd6969-6srgk_c21d8a72-9113-4b40-9cd2-fa7f56f749cd/neutron-api/0.log" Oct 11 04:24:40 crc kubenswrapper[4967]: I1011 04:24:40.373366 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_neutron-5f96fd6969-6srgk_c21d8a72-9113-4b40-9cd2-fa7f56f749cd/neutron-httpd/0.log" Oct 11 04:24:40 crc kubenswrapper[4967]: I1011 04:24:40.684951 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_4c3cc259-3802-4172-8ae6-446cfbcd4ca5/nova-api-log/0.log" Oct 11 04:24:40 crc kubenswrapper[4967]: I1011 04:24:40.775922 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-api-0_4c3cc259-3802-4172-8ae6-446cfbcd4ca5/nova-api-api/0.log" Oct 11 04:24:40 crc kubenswrapper[4967]: I1011 04:24:40.817264 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell0-conductor-0_67b30c76-971c-4f10-a692-4bbf2a1e8531/nova-cell0-conductor-conductor/0.log" Oct 11 04:24:41 crc kubenswrapper[4967]: I1011 04:24:41.036799 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-conductor-0_a9b2e413-f97a-4b67-9071-d1aed0db806f/nova-cell1-conductor-conductor/0.log" Oct 11 04:24:41 crc kubenswrapper[4967]: I1011 04:24:41.145602 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-cell1-novncproxy-0_2914dd42-ea01-44b5-a8cd-c23fc07e3170/nova-cell1-novncproxy-novncproxy/0.log" Oct 11 04:24:41 crc kubenswrapper[4967]: I1011 04:24:41.386024 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_8e701cd5-1cf9-4ed0-87c6-2023c41e7982/nova-metadata-log/0.log" Oct 11 04:24:41 crc kubenswrapper[4967]: I1011 04:24:41.571488 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-scheduler-0_388b5da4-c3cd-43f3-b839-cfd0b6c2a8f8/nova-scheduler-scheduler/0.log" Oct 11 04:24:41 crc kubenswrapper[4967]: I1011 04:24:41.616626 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_2f52a524-225c-4609-8cb0-0ec800e14aad/mysql-bootstrap/0.log" Oct 11 04:24:41 crc kubenswrapper[4967]: I1011 04:24:41.675778 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_nova-metadata-0_8e701cd5-1cf9-4ed0-87c6-2023c41e7982/nova-metadata-metadata/0.log" Oct 11 04:24:41 crc kubenswrapper[4967]: I1011 04:24:41.764921 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_2f52a524-225c-4609-8cb0-0ec800e14aad/galera/0.log" Oct 11 04:24:41 crc kubenswrapper[4967]: I1011 04:24:41.785188 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-cell1-galera-0_2f52a524-225c-4609-8cb0-0ec800e14aad/mysql-bootstrap/0.log" Oct 11 04:24:41 crc kubenswrapper[4967]: I1011 04:24:41.919046 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ddc18aa7-8599-40c5-a019-972c13162f83/mysql-bootstrap/0.log" Oct 11 04:24:42 crc kubenswrapper[4967]: I1011 04:24:42.082421 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ddc18aa7-8599-40c5-a019-972c13162f83/mysql-bootstrap/0.log" Oct 11 04:24:42 crc kubenswrapper[4967]: I1011 04:24:42.120971 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstack-galera-0_ddc18aa7-8599-40c5-a019-972c13162f83/galera/0.log" Oct 11 04:24:42 crc kubenswrapper[4967]: I1011 04:24:42.193242 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_openstackclient_23c5100e-07d0-4897-98de-f4e91329085e/openstackclient/0.log" Oct 11 04:24:42 crc kubenswrapper[4967]: I1011 04:24:42.337329 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-metrics-2n27p_6df16c29-dfa6-4972-9b1c-c1190bb1ca2a/openstack-network-exporter/0.log" Oct 11 04:24:42 crc kubenswrapper[4967]: I1011 04:24:42.409659 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7b5rp_4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20/ovsdb-server-init/0.log" Oct 11 04:24:42 crc kubenswrapper[4967]: I1011 04:24:42.640130 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7b5rp_4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20/ovsdb-server-init/0.log" Oct 11 04:24:42 crc kubenswrapper[4967]: I1011 04:24:42.651768 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7b5rp_4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20/ovs-vswitchd/0.log" Oct 11 04:24:42 crc kubenswrapper[4967]: I1011 04:24:42.657497 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-ovs-7b5rp_4ac64fc5-2a55-4d4e-bae7-9dcdf43c8a20/ovsdb-server/0.log" Oct 11 04:24:42 crc kubenswrapper[4967]: I1011 04:24:42.825346 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_942f72f2-4ef8-4056-b470-cbc9b2741f10/openstack-network-exporter/0.log" Oct 11 04:24:42 crc kubenswrapper[4967]: I1011 04:24:42.827254 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-controller-wvrnq_acad8def-db3f-40a3-a95d-28b7503d4aae/ovn-controller/0.log" Oct 11 04:24:42 crc kubenswrapper[4967]: I1011 04:24:42.916018 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovn-northd-0_942f72f2-4ef8-4056-b470-cbc9b2741f10/ovn-northd/0.log" Oct 11 04:24:43 crc kubenswrapper[4967]: I1011 04:24:43.040735 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_90a0e0c5-70a6-4330-986c-a09d244f781b/openstack-network-exporter/0.log" Oct 11 04:24:43 crc kubenswrapper[4967]: I1011 04:24:43.099939 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-nb-0_90a0e0c5-70a6-4330-986c-a09d244f781b/ovsdbserver-nb/0.log" Oct 11 04:24:43 crc kubenswrapper[4967]: I1011 04:24:43.237058 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca/openstack-network-exporter/0.log" Oct 11 04:24:43 crc kubenswrapper[4967]: I1011 04:24:43.293291 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_ovsdbserver-sb-0_36287cb5-e5d9-4fc1-9d8f-a36c4e5803ca/ovsdbserver-sb/0.log" Oct 11 04:24:43 crc kubenswrapper[4967]: I1011 04:24:43.415933 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5cdcf9b996-625cn_878b0a76-9c37-4210-870b-5fed33721831/placement-api/0.log" Oct 11 04:24:43 crc kubenswrapper[4967]: I1011 04:24:43.513729 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_c41a3ead-949a-4f76-9a58-0e1b4db0abb5/setup-container/0.log" Oct 11 04:24:43 crc kubenswrapper[4967]: I1011 04:24:43.548997 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_placement-5cdcf9b996-625cn_878b0a76-9c37-4210-870b-5fed33721831/placement-log/0.log" Oct 11 04:24:43 crc kubenswrapper[4967]: I1011 04:24:43.762114 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_c41a3ead-949a-4f76-9a58-0e1b4db0abb5/setup-container/0.log" Oct 11 04:24:43 crc kubenswrapper[4967]: I1011 04:24:43.770107 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-cell1-server-0_c41a3ead-949a-4f76-9a58-0e1b4db0abb5/rabbitmq/0.log" Oct 11 04:24:43 crc kubenswrapper[4967]: I1011 04:24:43.865482 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab/setup-container/0.log" Oct 11 04:24:44 crc kubenswrapper[4967]: I1011 04:24:44.052529 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab/setup-container/0.log" Oct 11 04:24:44 crc kubenswrapper[4967]: I1011 04:24:44.092831 4967 patch_prober.go:28] interesting pod/machine-config-daemon-sgrv2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:24:44 crc kubenswrapper[4967]: I1011 04:24:44.093143 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:24:44 crc kubenswrapper[4967]: I1011 04:24:44.093297 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" Oct 11 04:24:44 crc kubenswrapper[4967]: I1011 04:24:44.094266 4967 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d6d902b7233db28561ab70d9cf62192d380aef02bd4e78fc79f31d3c320df63c"} pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 11 04:24:44 crc kubenswrapper[4967]: I1011 04:24:44.094430 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" containerID="cri-o://d6d902b7233db28561ab70d9cf62192d380aef02bd4e78fc79f31d3c320df63c" gracePeriod=600 Oct 11 04:24:44 crc kubenswrapper[4967]: I1011 04:24:44.099489 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_rabbitmq-server-0_fb4c1b90-ae72-4a95-ac2f-1b70de0e10ab/rabbitmq/0.log" Oct 11 04:24:44 crc kubenswrapper[4967]: I1011 04:24:44.157604 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6486584599-dfqxj_fe5d2413-57a2-449d-9973-642e00e8c49f/proxy-httpd/0.log" Oct 11 04:24:44 crc kubenswrapper[4967]: I1011 04:24:44.268685 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-proxy-6486584599-dfqxj_fe5d2413-57a2-449d-9973-642e00e8c49f/proxy-server/0.log" Oct 11 04:24:44 crc kubenswrapper[4967]: I1011 04:24:44.412245 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-ring-rebalance-6rr4l_3f2085d6-0a42-4e1e-b52d-e9379d2a4838/swift-ring-rebalance/0.log" Oct 11 04:24:44 crc kubenswrapper[4967]: I1011 04:24:44.476103 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c09e3333-ad4d-495f-8973-7e2d2f0931d4/account-auditor/0.log" Oct 11 04:24:44 crc kubenswrapper[4967]: I1011 04:24:44.615606 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c09e3333-ad4d-495f-8973-7e2d2f0931d4/account-reaper/0.log" Oct 11 04:24:44 crc kubenswrapper[4967]: I1011 04:24:44.616193 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c09e3333-ad4d-495f-8973-7e2d2f0931d4/account-replicator/0.log" Oct 11 04:24:44 crc kubenswrapper[4967]: I1011 04:24:44.691216 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c09e3333-ad4d-495f-8973-7e2d2f0931d4/account-server/0.log" Oct 11 04:24:44 crc kubenswrapper[4967]: I1011 04:24:44.776721 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c09e3333-ad4d-495f-8973-7e2d2f0931d4/container-auditor/0.log" Oct 11 04:24:44 crc kubenswrapper[4967]: I1011 04:24:44.779502 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c09e3333-ad4d-495f-8973-7e2d2f0931d4/container-server/0.log" Oct 11 04:24:44 crc kubenswrapper[4967]: I1011 04:24:44.827810 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c09e3333-ad4d-495f-8973-7e2d2f0931d4/container-replicator/0.log" Oct 11 04:24:44 crc kubenswrapper[4967]: I1011 04:24:44.925684 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c09e3333-ad4d-495f-8973-7e2d2f0931d4/container-updater/0.log" Oct 11 04:24:45 crc kubenswrapper[4967]: I1011 04:24:45.014979 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c09e3333-ad4d-495f-8973-7e2d2f0931d4/object-replicator/0.log" Oct 11 04:24:45 crc kubenswrapper[4967]: I1011 04:24:45.018988 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c09e3333-ad4d-495f-8973-7e2d2f0931d4/object-auditor/0.log" Oct 11 04:24:45 crc kubenswrapper[4967]: I1011 04:24:45.019520 4967 generic.go:334] "Generic (PLEG): container finished" podID="97c07678-14be-410c-b61f-498cb49bc960" containerID="d6d902b7233db28561ab70d9cf62192d380aef02bd4e78fc79f31d3c320df63c" exitCode=0 Oct 11 04:24:45 crc kubenswrapper[4967]: I1011 04:24:45.019646 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" event={"ID":"97c07678-14be-410c-b61f-498cb49bc960","Type":"ContainerDied","Data":"d6d902b7233db28561ab70d9cf62192d380aef02bd4e78fc79f31d3c320df63c"} Oct 11 04:24:45 crc kubenswrapper[4967]: I1011 04:24:45.019740 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" event={"ID":"97c07678-14be-410c-b61f-498cb49bc960","Type":"ContainerStarted","Data":"dfd2cbbb285a882d3d3b0918a1f4268e32c2905a0b9d3bb49ba405a42cb0df28"} Oct 11 04:24:45 crc kubenswrapper[4967]: I1011 04:24:45.019839 4967 scope.go:117] "RemoveContainer" containerID="6010a3358387442fdc0ff1dfa8bf2d3fb285ef11cfb5f18ad38fe86548ca1f20" Oct 11 04:24:45 crc kubenswrapper[4967]: I1011 04:24:45.103662 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c09e3333-ad4d-495f-8973-7e2d2f0931d4/object-expirer/0.log" Oct 11 04:24:45 crc kubenswrapper[4967]: I1011 04:24:45.127949 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c09e3333-ad4d-495f-8973-7e2d2f0931d4/object-server/0.log" Oct 11 04:24:45 crc kubenswrapper[4967]: I1011 04:24:45.205547 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c09e3333-ad4d-495f-8973-7e2d2f0931d4/rsync/0.log" Oct 11 04:24:45 crc kubenswrapper[4967]: I1011 04:24:45.230105 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c09e3333-ad4d-495f-8973-7e2d2f0931d4/object-updater/0.log" Oct 11 04:24:45 crc kubenswrapper[4967]: I1011 04:24:45.291030 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_swift-storage-0_c09e3333-ad4d-495f-8973-7e2d2f0931d4/swift-recon-cron/0.log" Oct 11 04:24:48 crc kubenswrapper[4967]: I1011 04:24:48.547426 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack_memcached-0_10ea033f-281c-4519-8da7-93e598f08b11/memcached/0.log" Oct 11 04:25:06 crc kubenswrapper[4967]: I1011 04:25:06.849795 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk_9adf77e3-238d-4024-99db-ab21bdd51110/util/0.log" Oct 11 04:25:07 crc kubenswrapper[4967]: I1011 04:25:07.027520 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk_9adf77e3-238d-4024-99db-ab21bdd51110/pull/0.log" Oct 11 04:25:07 crc kubenswrapper[4967]: I1011 04:25:07.036995 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk_9adf77e3-238d-4024-99db-ab21bdd51110/pull/0.log" Oct 11 04:25:07 crc kubenswrapper[4967]: I1011 04:25:07.038210 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk_9adf77e3-238d-4024-99db-ab21bdd51110/util/0.log" Oct 11 04:25:07 crc kubenswrapper[4967]: I1011 04:25:07.170166 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk_9adf77e3-238d-4024-99db-ab21bdd51110/util/0.log" Oct 11 04:25:07 crc kubenswrapper[4967]: I1011 04:25:07.199434 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk_9adf77e3-238d-4024-99db-ab21bdd51110/extract/0.log" Oct 11 04:25:07 crc kubenswrapper[4967]: I1011 04:25:07.232788 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_8e7bfb5fb9eeef3a0afc04b2c68bdcfdc2de01908c97f1454948a802c9h8jbk_9adf77e3-238d-4024-99db-ab21bdd51110/pull/0.log" Oct 11 04:25:07 crc kubenswrapper[4967]: I1011 04:25:07.353018 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-xz92m_6f3fbb7d-1207-45a2-8213-c7c71bd703f1/kube-rbac-proxy/0.log" Oct 11 04:25:07 crc kubenswrapper[4967]: I1011 04:25:07.400598 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-5glgp_91f45294-2d2f-4b40-a1b4-a841c7748cc8/kube-rbac-proxy/0.log" Oct 11 04:25:07 crc kubenswrapper[4967]: I1011 04:25:07.402957 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_barbican-operator-controller-manager-64f84fcdbb-xz92m_6f3fbb7d-1207-45a2-8213-c7c71bd703f1/manager/0.log" Oct 11 04:25:07 crc kubenswrapper[4967]: I1011 04:25:07.559684 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_cinder-operator-controller-manager-59cdc64769-5glgp_91f45294-2d2f-4b40-a1b4-a841c7748cc8/manager/0.log" Oct 11 04:25:07 crc kubenswrapper[4967]: I1011 04:25:07.588172 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-9gn9l_2a2805ea-7d92-42e4-89cb-9b2672f72fc6/kube-rbac-proxy/0.log" Oct 11 04:25:07 crc kubenswrapper[4967]: I1011 04:25:07.619138 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_designate-operator-controller-manager-687df44cdb-9gn9l_2a2805ea-7d92-42e4-89cb-9b2672f72fc6/manager/0.log" Oct 11 04:25:07 crc kubenswrapper[4967]: I1011 04:25:07.813877 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-zt44b_44eb39fa-5b20-4c02-b039-f23d7c5220d2/kube-rbac-proxy/0.log" Oct 11 04:25:07 crc kubenswrapper[4967]: I1011 04:25:07.887695 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_glance-operator-controller-manager-7bb46cd7d-zt44b_44eb39fa-5b20-4c02-b039-f23d7c5220d2/manager/0.log" Oct 11 04:25:07 crc kubenswrapper[4967]: I1011 04:25:07.975022 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-ww7nt_ecd905f3-adb2-47ae-a86d-6fc0b4907a76/kube-rbac-proxy/0.log" Oct 11 04:25:08 crc kubenswrapper[4967]: I1011 04:25:08.000904 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_heat-operator-controller-manager-6d9967f8dd-ww7nt_ecd905f3-adb2-47ae-a86d-6fc0b4907a76/manager/0.log" Oct 11 04:25:08 crc kubenswrapper[4967]: I1011 04:25:08.037026 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-8ccrl_e9feb54d-dbfa-40ec-ac45-535cca431024/kube-rbac-proxy/0.log" Oct 11 04:25:08 crc kubenswrapper[4967]: I1011 04:25:08.146900 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_horizon-operator-controller-manager-6d74794d9b-8ccrl_e9feb54d-dbfa-40ec-ac45-535cca431024/manager/0.log" Oct 11 04:25:08 crc kubenswrapper[4967]: I1011 04:25:08.195501 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-685c45897-5gbwb_6628a0ae-2a55-48b0-a5f1-0439c7d0c273/kube-rbac-proxy/0.log" Oct 11 04:25:08 crc kubenswrapper[4967]: I1011 04:25:08.374651 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_infra-operator-controller-manager-685c45897-5gbwb_6628a0ae-2a55-48b0-a5f1-0439c7d0c273/manager/0.log" Oct 11 04:25:08 crc kubenswrapper[4967]: I1011 04:25:08.389900 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67494cb947-sklz6_674852f1-9729-4a6e-8c90-31741fa69b49/kube-rbac-proxy/0.log" Oct 11 04:25:08 crc kubenswrapper[4967]: I1011 04:25:08.456819 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ironic-operator-controller-manager-67494cb947-sklz6_674852f1-9729-4a6e-8c90-31741fa69b49/manager/0.log" Oct 11 04:25:08 crc kubenswrapper[4967]: I1011 04:25:08.576095 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-ddb98f99b-brbrs_950d3635-f160-45a1-9a4c-749720e662dc/kube-rbac-proxy/0.log" Oct 11 04:25:08 crc kubenswrapper[4967]: I1011 04:25:08.641514 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_keystone-operator-controller-manager-ddb98f99b-brbrs_950d3635-f160-45a1-9a4c-749720e662dc/manager/0.log" Oct 11 04:25:08 crc kubenswrapper[4967]: I1011 04:25:08.704122 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-ls9k6_20584c10-d8a1-41ca-a0a2-9465fa5190f2/kube-rbac-proxy/0.log" Oct 11 04:25:08 crc kubenswrapper[4967]: I1011 04:25:08.756751 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_manila-operator-controller-manager-59578bc799-ls9k6_20584c10-d8a1-41ca-a0a2-9465fa5190f2/manager/0.log" Oct 11 04:25:08 crc kubenswrapper[4967]: I1011 04:25:08.896490 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-fjr8k_ef8ae8f4-ad95-4f1f-bc8c-170fe5058dc9/kube-rbac-proxy/0.log" Oct 11 04:25:08 crc kubenswrapper[4967]: I1011 04:25:08.940930 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_mariadb-operator-controller-manager-5777b4f897-fjr8k_ef8ae8f4-ad95-4f1f-bc8c-170fe5058dc9/manager/0.log" Oct 11 04:25:09 crc kubenswrapper[4967]: I1011 04:25:09.094438 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-6zmv5_538ebbcb-d51c-472c-a2b4-80236b0c8c08/kube-rbac-proxy/0.log" Oct 11 04:25:09 crc kubenswrapper[4967]: I1011 04:25:09.162541 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_neutron-operator-controller-manager-797d478b46-6zmv5_538ebbcb-d51c-472c-a2b4-80236b0c8c08/manager/0.log" Oct 11 04:25:09 crc kubenswrapper[4967]: I1011 04:25:09.258488 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-59px8_7140d0e1-9937-491b-aa8a-16db830aff11/kube-rbac-proxy/0.log" Oct 11 04:25:09 crc kubenswrapper[4967]: I1011 04:25:09.364214 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_nova-operator-controller-manager-57bb74c7bf-59px8_7140d0e1-9937-491b-aa8a-16db830aff11/manager/0.log" Oct 11 04:25:09 crc kubenswrapper[4967]: I1011 04:25:09.388341 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-7clg2_7e917a6a-9720-4427-b079-33d112a981a3/kube-rbac-proxy/0.log" Oct 11 04:25:09 crc kubenswrapper[4967]: I1011 04:25:09.461016 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_octavia-operator-controller-manager-6d7c7ddf95-7clg2_7e917a6a-9720-4427-b079-33d112a981a3/manager/0.log" Oct 11 04:25:09 crc kubenswrapper[4967]: I1011 04:25:09.564504 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns_0cc11444-d341-4821-a743-6b4af1466e4a/kube-rbac-proxy/0.log" Oct 11 04:25:09 crc kubenswrapper[4967]: I1011 04:25:09.571920 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-baremetal-operator-controller-manager-6cc7fb757dbb7ns_0cc11444-d341-4821-a743-6b4af1466e4a/manager/0.log" Oct 11 04:25:09 crc kubenswrapper[4967]: I1011 04:25:09.731776 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-66664b49c9-tk2lg_620877eb-6f9c-4647-b545-1cca7a438690/kube-rbac-proxy/0.log" Oct 11 04:25:09 crc kubenswrapper[4967]: I1011 04:25:09.881525 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-84859cd6c-hlmz6_e36d39c8-9b9a-4564-b779-0c038ba68dfb/kube-rbac-proxy/0.log" Oct 11 04:25:10 crc kubenswrapper[4967]: I1011 04:25:10.081822 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-index-xvswq_3dafc1e8-5bc6-4b6b-baa4-8c8204d98d31/registry-server/0.log" Oct 11 04:25:10 crc kubenswrapper[4967]: I1011 04:25:10.119534 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-operator-84859cd6c-hlmz6_e36d39c8-9b9a-4564-b779-0c038ba68dfb/operator/0.log" Oct 11 04:25:10 crc kubenswrapper[4967]: I1011 04:25:10.289861 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-869cc7797f-kkrkc_c9e29e65-cdd1-408d-bed0-bc283147b283/kube-rbac-proxy/0.log" Oct 11 04:25:10 crc kubenswrapper[4967]: I1011 04:25:10.315111 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-vng4c_26b9d5a0-dfbe-4e53-aa03-4fbb4682019b/kube-rbac-proxy/0.log" Oct 11 04:25:10 crc kubenswrapper[4967]: I1011 04:25:10.355726 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_ovn-operator-controller-manager-869cc7797f-kkrkc_c9e29e65-cdd1-408d-bed0-bc283147b283/manager/0.log" Oct 11 04:25:10 crc kubenswrapper[4967]: I1011 04:25:10.467777 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_openstack-operator-controller-manager-66664b49c9-tk2lg_620877eb-6f9c-4647-b545-1cca7a438690/manager/0.log" Oct 11 04:25:10 crc kubenswrapper[4967]: I1011 04:25:10.535171 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_placement-operator-controller-manager-664664cb68-vng4c_26b9d5a0-dfbe-4e53-aa03-4fbb4682019b/manager/0.log" Oct 11 04:25:10 crc kubenswrapper[4967]: I1011 04:25:10.576643 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_rabbitmq-cluster-operator-manager-5f97d8c699-cmzk2_11e4f97c-6107-45fd-8eb1-e39ef7bed4ce/operator/0.log" Oct 11 04:25:10 crc kubenswrapper[4967]: I1011 04:25:10.791905 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-65knj_1094a8be-16bb-4986-b4a4-465d41880800/kube-rbac-proxy/0.log" Oct 11 04:25:10 crc kubenswrapper[4967]: I1011 04:25:10.799536 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_swift-operator-controller-manager-5f4d5dfdc6-65knj_1094a8be-16bb-4986-b4a4-465d41880800/manager/0.log" Oct 11 04:25:10 crc kubenswrapper[4967]: I1011 04:25:10.906407 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-578874c84d-fgz82_e82102ad-391b-4b29-ae47-dae18dba7743/kube-rbac-proxy/0.log" Oct 11 04:25:10 crc kubenswrapper[4967]: I1011 04:25:10.964581 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-ffcdd6c94-4s7hv_5e24d70e-8882-45c0-9f6b-e443549ecabb/kube-rbac-proxy/0.log" Oct 11 04:25:10 crc kubenswrapper[4967]: I1011 04:25:10.972871 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_telemetry-operator-controller-manager-578874c84d-fgz82_e82102ad-391b-4b29-ae47-dae18dba7743/manager/0.log" Oct 11 04:25:11 crc kubenswrapper[4967]: I1011 04:25:11.080859 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_test-operator-controller-manager-ffcdd6c94-4s7hv_5e24d70e-8882-45c0-9f6b-e443549ecabb/manager/0.log" Oct 11 04:25:11 crc kubenswrapper[4967]: I1011 04:25:11.149443 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-646675d848-zdpbn_10a0ea36-9248-4086-83e4-de8578492421/kube-rbac-proxy/0.log" Oct 11 04:25:11 crc kubenswrapper[4967]: I1011 04:25:11.185996 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openstack-operators_watcher-operator-controller-manager-646675d848-zdpbn_10a0ea36-9248-4086-83e4-de8578492421/manager/0.log" Oct 11 04:25:15 crc kubenswrapper[4967]: I1011 04:25:15.805144 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-8lbn9"] Oct 11 04:25:15 crc kubenswrapper[4967]: E1011 04:25:15.805944 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b8a9b80-ec80-4d85-bada-f0c1420e0f3f" containerName="container-00" Oct 11 04:25:15 crc kubenswrapper[4967]: I1011 04:25:15.805956 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b8a9b80-ec80-4d85-bada-f0c1420e0f3f" containerName="container-00" Oct 11 04:25:15 crc kubenswrapper[4967]: I1011 04:25:15.806174 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b8a9b80-ec80-4d85-bada-f0c1420e0f3f" containerName="container-00" Oct 11 04:25:15 crc kubenswrapper[4967]: I1011 04:25:15.807461 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8lbn9" Oct 11 04:25:15 crc kubenswrapper[4967]: I1011 04:25:15.819740 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8lbn9"] Oct 11 04:25:15 crc kubenswrapper[4967]: I1011 04:25:15.917509 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8be18f4f-56aa-4c11-9870-5d9e652338cd-catalog-content\") pod \"redhat-marketplace-8lbn9\" (UID: \"8be18f4f-56aa-4c11-9870-5d9e652338cd\") " pod="openshift-marketplace/redhat-marketplace-8lbn9" Oct 11 04:25:15 crc kubenswrapper[4967]: I1011 04:25:15.918114 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zt8t7\" (UniqueName: \"kubernetes.io/projected/8be18f4f-56aa-4c11-9870-5d9e652338cd-kube-api-access-zt8t7\") pod \"redhat-marketplace-8lbn9\" (UID: \"8be18f4f-56aa-4c11-9870-5d9e652338cd\") " pod="openshift-marketplace/redhat-marketplace-8lbn9" Oct 11 04:25:15 crc kubenswrapper[4967]: I1011 04:25:15.918439 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8be18f4f-56aa-4c11-9870-5d9e652338cd-utilities\") pod \"redhat-marketplace-8lbn9\" (UID: \"8be18f4f-56aa-4c11-9870-5d9e652338cd\") " pod="openshift-marketplace/redhat-marketplace-8lbn9" Oct 11 04:25:16 crc kubenswrapper[4967]: I1011 04:25:16.021531 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zt8t7\" (UniqueName: \"kubernetes.io/projected/8be18f4f-56aa-4c11-9870-5d9e652338cd-kube-api-access-zt8t7\") pod \"redhat-marketplace-8lbn9\" (UID: \"8be18f4f-56aa-4c11-9870-5d9e652338cd\") " pod="openshift-marketplace/redhat-marketplace-8lbn9" Oct 11 04:25:16 crc kubenswrapper[4967]: I1011 04:25:16.021979 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8be18f4f-56aa-4c11-9870-5d9e652338cd-utilities\") pod \"redhat-marketplace-8lbn9\" (UID: \"8be18f4f-56aa-4c11-9870-5d9e652338cd\") " pod="openshift-marketplace/redhat-marketplace-8lbn9" Oct 11 04:25:16 crc kubenswrapper[4967]: I1011 04:25:16.022103 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8be18f4f-56aa-4c11-9870-5d9e652338cd-catalog-content\") pod \"redhat-marketplace-8lbn9\" (UID: \"8be18f4f-56aa-4c11-9870-5d9e652338cd\") " pod="openshift-marketplace/redhat-marketplace-8lbn9" Oct 11 04:25:16 crc kubenswrapper[4967]: I1011 04:25:16.022534 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8be18f4f-56aa-4c11-9870-5d9e652338cd-utilities\") pod \"redhat-marketplace-8lbn9\" (UID: \"8be18f4f-56aa-4c11-9870-5d9e652338cd\") " pod="openshift-marketplace/redhat-marketplace-8lbn9" Oct 11 04:25:16 crc kubenswrapper[4967]: I1011 04:25:16.022607 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8be18f4f-56aa-4c11-9870-5d9e652338cd-catalog-content\") pod \"redhat-marketplace-8lbn9\" (UID: \"8be18f4f-56aa-4c11-9870-5d9e652338cd\") " pod="openshift-marketplace/redhat-marketplace-8lbn9" Oct 11 04:25:16 crc kubenswrapper[4967]: I1011 04:25:16.045192 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zt8t7\" (UniqueName: \"kubernetes.io/projected/8be18f4f-56aa-4c11-9870-5d9e652338cd-kube-api-access-zt8t7\") pod \"redhat-marketplace-8lbn9\" (UID: \"8be18f4f-56aa-4c11-9870-5d9e652338cd\") " pod="openshift-marketplace/redhat-marketplace-8lbn9" Oct 11 04:25:16 crc kubenswrapper[4967]: I1011 04:25:16.139606 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8lbn9" Oct 11 04:25:16 crc kubenswrapper[4967]: I1011 04:25:16.628557 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-8lbn9"] Oct 11 04:25:17 crc kubenswrapper[4967]: I1011 04:25:17.305436 4967 generic.go:334] "Generic (PLEG): container finished" podID="8be18f4f-56aa-4c11-9870-5d9e652338cd" containerID="33e56714ffc30bb926edfcc705bf790c88fee1b11323efb989449eacd73db54a" exitCode=0 Oct 11 04:25:17 crc kubenswrapper[4967]: I1011 04:25:17.305529 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8lbn9" event={"ID":"8be18f4f-56aa-4c11-9870-5d9e652338cd","Type":"ContainerDied","Data":"33e56714ffc30bb926edfcc705bf790c88fee1b11323efb989449eacd73db54a"} Oct 11 04:25:17 crc kubenswrapper[4967]: I1011 04:25:17.305786 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8lbn9" event={"ID":"8be18f4f-56aa-4c11-9870-5d9e652338cd","Type":"ContainerStarted","Data":"8af65d54760addda253f750da475507ed42850b1a119354fd527930337dfdd2d"} Oct 11 04:25:17 crc kubenswrapper[4967]: I1011 04:25:17.307544 4967 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 11 04:25:18 crc kubenswrapper[4967]: I1011 04:25:18.015519 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-k454s"] Oct 11 04:25:18 crc kubenswrapper[4967]: I1011 04:25:18.020216 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k454s" Oct 11 04:25:18 crc kubenswrapper[4967]: I1011 04:25:18.029051 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k454s"] Oct 11 04:25:18 crc kubenswrapper[4967]: I1011 04:25:18.162587 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd082bfe-1c3a-4801-9ff2-dbe7e47740d1-catalog-content\") pod \"community-operators-k454s\" (UID: \"dd082bfe-1c3a-4801-9ff2-dbe7e47740d1\") " pod="openshift-marketplace/community-operators-k454s" Oct 11 04:25:18 crc kubenswrapper[4967]: I1011 04:25:18.162670 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x6q24\" (UniqueName: \"kubernetes.io/projected/dd082bfe-1c3a-4801-9ff2-dbe7e47740d1-kube-api-access-x6q24\") pod \"community-operators-k454s\" (UID: \"dd082bfe-1c3a-4801-9ff2-dbe7e47740d1\") " pod="openshift-marketplace/community-operators-k454s" Oct 11 04:25:18 crc kubenswrapper[4967]: I1011 04:25:18.162761 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd082bfe-1c3a-4801-9ff2-dbe7e47740d1-utilities\") pod \"community-operators-k454s\" (UID: \"dd082bfe-1c3a-4801-9ff2-dbe7e47740d1\") " pod="openshift-marketplace/community-operators-k454s" Oct 11 04:25:18 crc kubenswrapper[4967]: I1011 04:25:18.199486 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-h29dw"] Oct 11 04:25:18 crc kubenswrapper[4967]: I1011 04:25:18.201964 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h29dw" Oct 11 04:25:18 crc kubenswrapper[4967]: I1011 04:25:18.209786 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h29dw"] Oct 11 04:25:18 crc kubenswrapper[4967]: I1011 04:25:18.265366 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd082bfe-1c3a-4801-9ff2-dbe7e47740d1-catalog-content\") pod \"community-operators-k454s\" (UID: \"dd082bfe-1c3a-4801-9ff2-dbe7e47740d1\") " pod="openshift-marketplace/community-operators-k454s" Oct 11 04:25:18 crc kubenswrapper[4967]: I1011 04:25:18.265451 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x6q24\" (UniqueName: \"kubernetes.io/projected/dd082bfe-1c3a-4801-9ff2-dbe7e47740d1-kube-api-access-x6q24\") pod \"community-operators-k454s\" (UID: \"dd082bfe-1c3a-4801-9ff2-dbe7e47740d1\") " pod="openshift-marketplace/community-operators-k454s" Oct 11 04:25:18 crc kubenswrapper[4967]: I1011 04:25:18.265519 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd082bfe-1c3a-4801-9ff2-dbe7e47740d1-utilities\") pod \"community-operators-k454s\" (UID: \"dd082bfe-1c3a-4801-9ff2-dbe7e47740d1\") " pod="openshift-marketplace/community-operators-k454s" Oct 11 04:25:18 crc kubenswrapper[4967]: I1011 04:25:18.266134 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd082bfe-1c3a-4801-9ff2-dbe7e47740d1-catalog-content\") pod \"community-operators-k454s\" (UID: \"dd082bfe-1c3a-4801-9ff2-dbe7e47740d1\") " pod="openshift-marketplace/community-operators-k454s" Oct 11 04:25:18 crc kubenswrapper[4967]: I1011 04:25:18.266482 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd082bfe-1c3a-4801-9ff2-dbe7e47740d1-utilities\") pod \"community-operators-k454s\" (UID: \"dd082bfe-1c3a-4801-9ff2-dbe7e47740d1\") " pod="openshift-marketplace/community-operators-k454s" Oct 11 04:25:18 crc kubenswrapper[4967]: I1011 04:25:18.287745 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x6q24\" (UniqueName: \"kubernetes.io/projected/dd082bfe-1c3a-4801-9ff2-dbe7e47740d1-kube-api-access-x6q24\") pod \"community-operators-k454s\" (UID: \"dd082bfe-1c3a-4801-9ff2-dbe7e47740d1\") " pod="openshift-marketplace/community-operators-k454s" Oct 11 04:25:18 crc kubenswrapper[4967]: I1011 04:25:18.315390 4967 generic.go:334] "Generic (PLEG): container finished" podID="8be18f4f-56aa-4c11-9870-5d9e652338cd" containerID="23669c05fde48ad2d48557c7c55eb6839d3f1e95463559ee4e3da0acae81c3c1" exitCode=0 Oct 11 04:25:18 crc kubenswrapper[4967]: I1011 04:25:18.315430 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8lbn9" event={"ID":"8be18f4f-56aa-4c11-9870-5d9e652338cd","Type":"ContainerDied","Data":"23669c05fde48ad2d48557c7c55eb6839d3f1e95463559ee4e3da0acae81c3c1"} Oct 11 04:25:18 crc kubenswrapper[4967]: I1011 04:25:18.367145 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-z6r2p\" (UniqueName: \"kubernetes.io/projected/343b1a7b-fa49-4f53-ae16-2a827ed81936-kube-api-access-z6r2p\") pod \"redhat-operators-h29dw\" (UID: \"343b1a7b-fa49-4f53-ae16-2a827ed81936\") " pod="openshift-marketplace/redhat-operators-h29dw" Oct 11 04:25:18 crc kubenswrapper[4967]: I1011 04:25:18.367399 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/343b1a7b-fa49-4f53-ae16-2a827ed81936-utilities\") pod \"redhat-operators-h29dw\" (UID: \"343b1a7b-fa49-4f53-ae16-2a827ed81936\") " pod="openshift-marketplace/redhat-operators-h29dw" Oct 11 04:25:18 crc kubenswrapper[4967]: I1011 04:25:18.368322 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/343b1a7b-fa49-4f53-ae16-2a827ed81936-catalog-content\") pod \"redhat-operators-h29dw\" (UID: \"343b1a7b-fa49-4f53-ae16-2a827ed81936\") " pod="openshift-marketplace/redhat-operators-h29dw" Oct 11 04:25:18 crc kubenswrapper[4967]: I1011 04:25:18.385330 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k454s" Oct 11 04:25:18 crc kubenswrapper[4967]: I1011 04:25:18.470103 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-z6r2p\" (UniqueName: \"kubernetes.io/projected/343b1a7b-fa49-4f53-ae16-2a827ed81936-kube-api-access-z6r2p\") pod \"redhat-operators-h29dw\" (UID: \"343b1a7b-fa49-4f53-ae16-2a827ed81936\") " pod="openshift-marketplace/redhat-operators-h29dw" Oct 11 04:25:18 crc kubenswrapper[4967]: I1011 04:25:18.470168 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/343b1a7b-fa49-4f53-ae16-2a827ed81936-utilities\") pod \"redhat-operators-h29dw\" (UID: \"343b1a7b-fa49-4f53-ae16-2a827ed81936\") " pod="openshift-marketplace/redhat-operators-h29dw" Oct 11 04:25:18 crc kubenswrapper[4967]: I1011 04:25:18.470245 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/343b1a7b-fa49-4f53-ae16-2a827ed81936-catalog-content\") pod \"redhat-operators-h29dw\" (UID: \"343b1a7b-fa49-4f53-ae16-2a827ed81936\") " pod="openshift-marketplace/redhat-operators-h29dw" Oct 11 04:25:18 crc kubenswrapper[4967]: I1011 04:25:18.470803 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/343b1a7b-fa49-4f53-ae16-2a827ed81936-catalog-content\") pod \"redhat-operators-h29dw\" (UID: \"343b1a7b-fa49-4f53-ae16-2a827ed81936\") " pod="openshift-marketplace/redhat-operators-h29dw" Oct 11 04:25:18 crc kubenswrapper[4967]: I1011 04:25:18.470865 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/343b1a7b-fa49-4f53-ae16-2a827ed81936-utilities\") pod \"redhat-operators-h29dw\" (UID: \"343b1a7b-fa49-4f53-ae16-2a827ed81936\") " pod="openshift-marketplace/redhat-operators-h29dw" Oct 11 04:25:18 crc kubenswrapper[4967]: I1011 04:25:18.487432 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-z6r2p\" (UniqueName: \"kubernetes.io/projected/343b1a7b-fa49-4f53-ae16-2a827ed81936-kube-api-access-z6r2p\") pod \"redhat-operators-h29dw\" (UID: \"343b1a7b-fa49-4f53-ae16-2a827ed81936\") " pod="openshift-marketplace/redhat-operators-h29dw" Oct 11 04:25:18 crc kubenswrapper[4967]: I1011 04:25:18.519861 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h29dw" Oct 11 04:25:18 crc kubenswrapper[4967]: I1011 04:25:18.961609 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-k454s"] Oct 11 04:25:18 crc kubenswrapper[4967]: W1011 04:25:18.965490 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poddd082bfe_1c3a_4801_9ff2_dbe7e47740d1.slice/crio-344ac7c54888bc157c2c850a62d71b2c4caff85abc9c47de85322cf05d3a5fbd WatchSource:0}: Error finding container 344ac7c54888bc157c2c850a62d71b2c4caff85abc9c47de85322cf05d3a5fbd: Status 404 returned error can't find the container with id 344ac7c54888bc157c2c850a62d71b2c4caff85abc9c47de85322cf05d3a5fbd Oct 11 04:25:19 crc kubenswrapper[4967]: I1011 04:25:19.049734 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-h29dw"] Oct 11 04:25:19 crc kubenswrapper[4967]: W1011 04:25:19.055856 4967 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod343b1a7b_fa49_4f53_ae16_2a827ed81936.slice/crio-72292d36e85690940039f619eba12d6b108183f7bffb28ede608a6faa10f6250 WatchSource:0}: Error finding container 72292d36e85690940039f619eba12d6b108183f7bffb28ede608a6faa10f6250: Status 404 returned error can't find the container with id 72292d36e85690940039f619eba12d6b108183f7bffb28ede608a6faa10f6250 Oct 11 04:25:19 crc kubenswrapper[4967]: I1011 04:25:19.324490 4967 generic.go:334] "Generic (PLEG): container finished" podID="343b1a7b-fa49-4f53-ae16-2a827ed81936" containerID="f95d1f5d07e67406dcf99f3c920a25533d38388bc349356588e1b6790ec4b68e" exitCode=0 Oct 11 04:25:19 crc kubenswrapper[4967]: I1011 04:25:19.324528 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h29dw" event={"ID":"343b1a7b-fa49-4f53-ae16-2a827ed81936","Type":"ContainerDied","Data":"f95d1f5d07e67406dcf99f3c920a25533d38388bc349356588e1b6790ec4b68e"} Oct 11 04:25:19 crc kubenswrapper[4967]: I1011 04:25:19.324821 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h29dw" event={"ID":"343b1a7b-fa49-4f53-ae16-2a827ed81936","Type":"ContainerStarted","Data":"72292d36e85690940039f619eba12d6b108183f7bffb28ede608a6faa10f6250"} Oct 11 04:25:19 crc kubenswrapper[4967]: I1011 04:25:19.327623 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8lbn9" event={"ID":"8be18f4f-56aa-4c11-9870-5d9e652338cd","Type":"ContainerStarted","Data":"bbaee0bc53b31ea68cca69712e030de7124d1b3f094e9a1eac97ca40678aca9f"} Oct 11 04:25:19 crc kubenswrapper[4967]: I1011 04:25:19.334580 4967 generic.go:334] "Generic (PLEG): container finished" podID="dd082bfe-1c3a-4801-9ff2-dbe7e47740d1" containerID="0edc7ef82b8f206fed69ef8bde52a262f731fc1414ad1dd6f00d861ef630a5c8" exitCode=0 Oct 11 04:25:19 crc kubenswrapper[4967]: I1011 04:25:19.334623 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k454s" event={"ID":"dd082bfe-1c3a-4801-9ff2-dbe7e47740d1","Type":"ContainerDied","Data":"0edc7ef82b8f206fed69ef8bde52a262f731fc1414ad1dd6f00d861ef630a5c8"} Oct 11 04:25:19 crc kubenswrapper[4967]: I1011 04:25:19.334648 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k454s" event={"ID":"dd082bfe-1c3a-4801-9ff2-dbe7e47740d1","Type":"ContainerStarted","Data":"344ac7c54888bc157c2c850a62d71b2c4caff85abc9c47de85322cf05d3a5fbd"} Oct 11 04:25:19 crc kubenswrapper[4967]: I1011 04:25:19.378582 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-8lbn9" podStartSLOduration=2.965836725 podStartE2EDuration="4.378561889s" podCreationTimestamp="2025-10-11 04:25:15 +0000 UTC" firstStartedPulling="2025-10-11 04:25:17.307297505 +0000 UTC m=+2045.270506438" lastFinishedPulling="2025-10-11 04:25:18.720022669 +0000 UTC m=+2046.683231602" observedRunningTime="2025-10-11 04:25:19.369045458 +0000 UTC m=+2047.332254391" watchObservedRunningTime="2025-10-11 04:25:19.378561889 +0000 UTC m=+2047.341770822" Oct 11 04:25:20 crc kubenswrapper[4967]: I1011 04:25:20.346200 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h29dw" event={"ID":"343b1a7b-fa49-4f53-ae16-2a827ed81936","Type":"ContainerStarted","Data":"1da7b05790ce1a95ae995901cc3ba6ad3dace5e6f1468b56fd12ffd92139fbbe"} Oct 11 04:25:21 crc kubenswrapper[4967]: I1011 04:25:21.364394 4967 generic.go:334] "Generic (PLEG): container finished" podID="dd082bfe-1c3a-4801-9ff2-dbe7e47740d1" containerID="52baddc35a310d149c81b0138b4df19d53374f55b9de280f7ff41ff8595f55b7" exitCode=0 Oct 11 04:25:21 crc kubenswrapper[4967]: I1011 04:25:21.366342 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k454s" event={"ID":"dd082bfe-1c3a-4801-9ff2-dbe7e47740d1","Type":"ContainerDied","Data":"52baddc35a310d149c81b0138b4df19d53374f55b9de280f7ff41ff8595f55b7"} Oct 11 04:25:22 crc kubenswrapper[4967]: I1011 04:25:22.375163 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k454s" event={"ID":"dd082bfe-1c3a-4801-9ff2-dbe7e47740d1","Type":"ContainerStarted","Data":"326cf2e8aff7bd2982619535424d0d2e23e8c8031fbf0879b0520e83419671a6"} Oct 11 04:25:22 crc kubenswrapper[4967]: I1011 04:25:22.397253 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-k454s" podStartSLOduration=2.947365602 podStartE2EDuration="5.397230332s" podCreationTimestamp="2025-10-11 04:25:17 +0000 UTC" firstStartedPulling="2025-10-11 04:25:19.339166303 +0000 UTC m=+2047.302375236" lastFinishedPulling="2025-10-11 04:25:21.789030993 +0000 UTC m=+2049.752239966" observedRunningTime="2025-10-11 04:25:22.393433063 +0000 UTC m=+2050.356641996" watchObservedRunningTime="2025-10-11 04:25:22.397230332 +0000 UTC m=+2050.360439305" Oct 11 04:25:23 crc kubenswrapper[4967]: I1011 04:25:23.386583 4967 generic.go:334] "Generic (PLEG): container finished" podID="343b1a7b-fa49-4f53-ae16-2a827ed81936" containerID="1da7b05790ce1a95ae995901cc3ba6ad3dace5e6f1468b56fd12ffd92139fbbe" exitCode=0 Oct 11 04:25:23 crc kubenswrapper[4967]: I1011 04:25:23.386705 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h29dw" event={"ID":"343b1a7b-fa49-4f53-ae16-2a827ed81936","Type":"ContainerDied","Data":"1da7b05790ce1a95ae995901cc3ba6ad3dace5e6f1468b56fd12ffd92139fbbe"} Oct 11 04:25:24 crc kubenswrapper[4967]: I1011 04:25:24.398245 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h29dw" event={"ID":"343b1a7b-fa49-4f53-ae16-2a827ed81936","Type":"ContainerStarted","Data":"94e5bfb7ca7b0e033e3d140b2e30e2cc85b331459f8f73bb61685d098542457b"} Oct 11 04:25:24 crc kubenswrapper[4967]: I1011 04:25:24.419343 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-h29dw" podStartSLOduration=1.9303107050000001 podStartE2EDuration="6.419326181s" podCreationTimestamp="2025-10-11 04:25:18 +0000 UTC" firstStartedPulling="2025-10-11 04:25:19.325976906 +0000 UTC m=+2047.289185839" lastFinishedPulling="2025-10-11 04:25:23.814992382 +0000 UTC m=+2051.778201315" observedRunningTime="2025-10-11 04:25:24.412919442 +0000 UTC m=+2052.376128375" watchObservedRunningTime="2025-10-11 04:25:24.419326181 +0000 UTC m=+2052.382535114" Oct 11 04:25:25 crc kubenswrapper[4967]: I1011 04:25:25.755544 4967 scope.go:117] "RemoveContainer" containerID="7a838cf131945ab75399cf021780705d1d3db35b4babc83434cd6d5b26cdd25a" Oct 11 04:25:25 crc kubenswrapper[4967]: I1011 04:25:25.783524 4967 scope.go:117] "RemoveContainer" containerID="aaf10915876775ab06c02796a259fdd767edb2dc8371c87aa795b922be5a13be" Oct 11 04:25:26 crc kubenswrapper[4967]: I1011 04:25:26.140210 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-8lbn9" Oct 11 04:25:26 crc kubenswrapper[4967]: I1011 04:25:26.140269 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-8lbn9" Oct 11 04:25:26 crc kubenswrapper[4967]: I1011 04:25:26.202893 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-8lbn9" Oct 11 04:25:26 crc kubenswrapper[4967]: I1011 04:25:26.464791 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-8lbn9" Oct 11 04:25:27 crc kubenswrapper[4967]: I1011 04:25:27.286671 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-rv6n2_c3a20ddb-cfaa-46ac-8aed-4db9de3e32b6/control-plane-machine-set-operator/0.log" Oct 11 04:25:27 crc kubenswrapper[4967]: I1011 04:25:27.498178 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-mrsjk_9b3efa80-8370-4353-b688-4701c4deec8f/kube-rbac-proxy/0.log" Oct 11 04:25:27 crc kubenswrapper[4967]: I1011 04:25:27.498935 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-mrsjk_9b3efa80-8370-4353-b688-4701c4deec8f/machine-api-operator/0.log" Oct 11 04:25:27 crc kubenswrapper[4967]: I1011 04:25:27.785139 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8lbn9"] Oct 11 04:25:28 crc kubenswrapper[4967]: I1011 04:25:28.385918 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-k454s" Oct 11 04:25:28 crc kubenswrapper[4967]: I1011 04:25:28.386616 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-k454s" Oct 11 04:25:28 crc kubenswrapper[4967]: I1011 04:25:28.429952 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-8lbn9" podUID="8be18f4f-56aa-4c11-9870-5d9e652338cd" containerName="registry-server" containerID="cri-o://bbaee0bc53b31ea68cca69712e030de7124d1b3f094e9a1eac97ca40678aca9f" gracePeriod=2 Oct 11 04:25:28 crc kubenswrapper[4967]: I1011 04:25:28.435973 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-k454s" Oct 11 04:25:28 crc kubenswrapper[4967]: I1011 04:25:28.490990 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-k454s" Oct 11 04:25:28 crc kubenswrapper[4967]: I1011 04:25:28.520557 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-h29dw" Oct 11 04:25:28 crc kubenswrapper[4967]: I1011 04:25:28.521011 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-h29dw" Oct 11 04:25:28 crc kubenswrapper[4967]: I1011 04:25:28.941250 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8lbn9" Oct 11 04:25:29 crc kubenswrapper[4967]: I1011 04:25:29.075190 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8be18f4f-56aa-4c11-9870-5d9e652338cd-utilities\") pod \"8be18f4f-56aa-4c11-9870-5d9e652338cd\" (UID: \"8be18f4f-56aa-4c11-9870-5d9e652338cd\") " Oct 11 04:25:29 crc kubenswrapper[4967]: I1011 04:25:29.075226 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zt8t7\" (UniqueName: \"kubernetes.io/projected/8be18f4f-56aa-4c11-9870-5d9e652338cd-kube-api-access-zt8t7\") pod \"8be18f4f-56aa-4c11-9870-5d9e652338cd\" (UID: \"8be18f4f-56aa-4c11-9870-5d9e652338cd\") " Oct 11 04:25:29 crc kubenswrapper[4967]: I1011 04:25:29.075361 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8be18f4f-56aa-4c11-9870-5d9e652338cd-catalog-content\") pod \"8be18f4f-56aa-4c11-9870-5d9e652338cd\" (UID: \"8be18f4f-56aa-4c11-9870-5d9e652338cd\") " Oct 11 04:25:29 crc kubenswrapper[4967]: I1011 04:25:29.075915 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8be18f4f-56aa-4c11-9870-5d9e652338cd-utilities" (OuterVolumeSpecName: "utilities") pod "8be18f4f-56aa-4c11-9870-5d9e652338cd" (UID: "8be18f4f-56aa-4c11-9870-5d9e652338cd"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:25:29 crc kubenswrapper[4967]: I1011 04:25:29.081649 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8be18f4f-56aa-4c11-9870-5d9e652338cd-kube-api-access-zt8t7" (OuterVolumeSpecName: "kube-api-access-zt8t7") pod "8be18f4f-56aa-4c11-9870-5d9e652338cd" (UID: "8be18f4f-56aa-4c11-9870-5d9e652338cd"). InnerVolumeSpecName "kube-api-access-zt8t7". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:25:29 crc kubenswrapper[4967]: I1011 04:25:29.087862 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8be18f4f-56aa-4c11-9870-5d9e652338cd-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8be18f4f-56aa-4c11-9870-5d9e652338cd" (UID: "8be18f4f-56aa-4c11-9870-5d9e652338cd"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:25:29 crc kubenswrapper[4967]: I1011 04:25:29.177157 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8be18f4f-56aa-4c11-9870-5d9e652338cd-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 04:25:29 crc kubenswrapper[4967]: I1011 04:25:29.177385 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zt8t7\" (UniqueName: \"kubernetes.io/projected/8be18f4f-56aa-4c11-9870-5d9e652338cd-kube-api-access-zt8t7\") on node \"crc\" DevicePath \"\"" Oct 11 04:25:29 crc kubenswrapper[4967]: I1011 04:25:29.177443 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8be18f4f-56aa-4c11-9870-5d9e652338cd-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 04:25:29 crc kubenswrapper[4967]: I1011 04:25:29.443016 4967 generic.go:334] "Generic (PLEG): container finished" podID="8be18f4f-56aa-4c11-9870-5d9e652338cd" containerID="bbaee0bc53b31ea68cca69712e030de7124d1b3f094e9a1eac97ca40678aca9f" exitCode=0 Oct 11 04:25:29 crc kubenswrapper[4967]: I1011 04:25:29.443098 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8lbn9" event={"ID":"8be18f4f-56aa-4c11-9870-5d9e652338cd","Type":"ContainerDied","Data":"bbaee0bc53b31ea68cca69712e030de7124d1b3f094e9a1eac97ca40678aca9f"} Oct 11 04:25:29 crc kubenswrapper[4967]: I1011 04:25:29.443158 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-8lbn9" Oct 11 04:25:29 crc kubenswrapper[4967]: I1011 04:25:29.443171 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-8lbn9" event={"ID":"8be18f4f-56aa-4c11-9870-5d9e652338cd","Type":"ContainerDied","Data":"8af65d54760addda253f750da475507ed42850b1a119354fd527930337dfdd2d"} Oct 11 04:25:29 crc kubenswrapper[4967]: I1011 04:25:29.443244 4967 scope.go:117] "RemoveContainer" containerID="bbaee0bc53b31ea68cca69712e030de7124d1b3f094e9a1eac97ca40678aca9f" Oct 11 04:25:29 crc kubenswrapper[4967]: I1011 04:25:29.480011 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-8lbn9"] Oct 11 04:25:29 crc kubenswrapper[4967]: I1011 04:25:29.485550 4967 scope.go:117] "RemoveContainer" containerID="23669c05fde48ad2d48557c7c55eb6839d3f1e95463559ee4e3da0acae81c3c1" Oct 11 04:25:29 crc kubenswrapper[4967]: I1011 04:25:29.491902 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-8lbn9"] Oct 11 04:25:29 crc kubenswrapper[4967]: I1011 04:25:29.507926 4967 scope.go:117] "RemoveContainer" containerID="33e56714ffc30bb926edfcc705bf790c88fee1b11323efb989449eacd73db54a" Oct 11 04:25:29 crc kubenswrapper[4967]: I1011 04:25:29.562457 4967 scope.go:117] "RemoveContainer" containerID="bbaee0bc53b31ea68cca69712e030de7124d1b3f094e9a1eac97ca40678aca9f" Oct 11 04:25:29 crc kubenswrapper[4967]: E1011 04:25:29.563085 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bbaee0bc53b31ea68cca69712e030de7124d1b3f094e9a1eac97ca40678aca9f\": container with ID starting with bbaee0bc53b31ea68cca69712e030de7124d1b3f094e9a1eac97ca40678aca9f not found: ID does not exist" containerID="bbaee0bc53b31ea68cca69712e030de7124d1b3f094e9a1eac97ca40678aca9f" Oct 11 04:25:29 crc kubenswrapper[4967]: I1011 04:25:29.563134 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bbaee0bc53b31ea68cca69712e030de7124d1b3f094e9a1eac97ca40678aca9f"} err="failed to get container status \"bbaee0bc53b31ea68cca69712e030de7124d1b3f094e9a1eac97ca40678aca9f\": rpc error: code = NotFound desc = could not find container \"bbaee0bc53b31ea68cca69712e030de7124d1b3f094e9a1eac97ca40678aca9f\": container with ID starting with bbaee0bc53b31ea68cca69712e030de7124d1b3f094e9a1eac97ca40678aca9f not found: ID does not exist" Oct 11 04:25:29 crc kubenswrapper[4967]: I1011 04:25:29.563160 4967 scope.go:117] "RemoveContainer" containerID="23669c05fde48ad2d48557c7c55eb6839d3f1e95463559ee4e3da0acae81c3c1" Oct 11 04:25:29 crc kubenswrapper[4967]: E1011 04:25:29.563563 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"23669c05fde48ad2d48557c7c55eb6839d3f1e95463559ee4e3da0acae81c3c1\": container with ID starting with 23669c05fde48ad2d48557c7c55eb6839d3f1e95463559ee4e3da0acae81c3c1 not found: ID does not exist" containerID="23669c05fde48ad2d48557c7c55eb6839d3f1e95463559ee4e3da0acae81c3c1" Oct 11 04:25:29 crc kubenswrapper[4967]: I1011 04:25:29.563615 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"23669c05fde48ad2d48557c7c55eb6839d3f1e95463559ee4e3da0acae81c3c1"} err="failed to get container status \"23669c05fde48ad2d48557c7c55eb6839d3f1e95463559ee4e3da0acae81c3c1\": rpc error: code = NotFound desc = could not find container \"23669c05fde48ad2d48557c7c55eb6839d3f1e95463559ee4e3da0acae81c3c1\": container with ID starting with 23669c05fde48ad2d48557c7c55eb6839d3f1e95463559ee4e3da0acae81c3c1 not found: ID does not exist" Oct 11 04:25:29 crc kubenswrapper[4967]: I1011 04:25:29.563646 4967 scope.go:117] "RemoveContainer" containerID="33e56714ffc30bb926edfcc705bf790c88fee1b11323efb989449eacd73db54a" Oct 11 04:25:29 crc kubenswrapper[4967]: E1011 04:25:29.564110 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"33e56714ffc30bb926edfcc705bf790c88fee1b11323efb989449eacd73db54a\": container with ID starting with 33e56714ffc30bb926edfcc705bf790c88fee1b11323efb989449eacd73db54a not found: ID does not exist" containerID="33e56714ffc30bb926edfcc705bf790c88fee1b11323efb989449eacd73db54a" Oct 11 04:25:29 crc kubenswrapper[4967]: I1011 04:25:29.564136 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"33e56714ffc30bb926edfcc705bf790c88fee1b11323efb989449eacd73db54a"} err="failed to get container status \"33e56714ffc30bb926edfcc705bf790c88fee1b11323efb989449eacd73db54a\": rpc error: code = NotFound desc = could not find container \"33e56714ffc30bb926edfcc705bf790c88fee1b11323efb989449eacd73db54a\": container with ID starting with 33e56714ffc30bb926edfcc705bf790c88fee1b11323efb989449eacd73db54a not found: ID does not exist" Oct 11 04:25:29 crc kubenswrapper[4967]: I1011 04:25:29.569059 4967 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-h29dw" podUID="343b1a7b-fa49-4f53-ae16-2a827ed81936" containerName="registry-server" probeResult="failure" output=< Oct 11 04:25:29 crc kubenswrapper[4967]: timeout: failed to connect service ":50051" within 1s Oct 11 04:25:29 crc kubenswrapper[4967]: > Oct 11 04:25:30 crc kubenswrapper[4967]: I1011 04:25:30.788430 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k454s"] Oct 11 04:25:30 crc kubenswrapper[4967]: I1011 04:25:30.789135 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-k454s" podUID="dd082bfe-1c3a-4801-9ff2-dbe7e47740d1" containerName="registry-server" containerID="cri-o://326cf2e8aff7bd2982619535424d0d2e23e8c8031fbf0879b0520e83419671a6" gracePeriod=2 Oct 11 04:25:30 crc kubenswrapper[4967]: I1011 04:25:30.826963 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8be18f4f-56aa-4c11-9870-5d9e652338cd" path="/var/lib/kubelet/pods/8be18f4f-56aa-4c11-9870-5d9e652338cd/volumes" Oct 11 04:25:31 crc kubenswrapper[4967]: I1011 04:25:31.273156 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k454s" Oct 11 04:25:31 crc kubenswrapper[4967]: I1011 04:25:31.322846 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd082bfe-1c3a-4801-9ff2-dbe7e47740d1-utilities\") pod \"dd082bfe-1c3a-4801-9ff2-dbe7e47740d1\" (UID: \"dd082bfe-1c3a-4801-9ff2-dbe7e47740d1\") " Oct 11 04:25:31 crc kubenswrapper[4967]: I1011 04:25:31.322937 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd082bfe-1c3a-4801-9ff2-dbe7e47740d1-catalog-content\") pod \"dd082bfe-1c3a-4801-9ff2-dbe7e47740d1\" (UID: \"dd082bfe-1c3a-4801-9ff2-dbe7e47740d1\") " Oct 11 04:25:31 crc kubenswrapper[4967]: I1011 04:25:31.323289 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x6q24\" (UniqueName: \"kubernetes.io/projected/dd082bfe-1c3a-4801-9ff2-dbe7e47740d1-kube-api-access-x6q24\") pod \"dd082bfe-1c3a-4801-9ff2-dbe7e47740d1\" (UID: \"dd082bfe-1c3a-4801-9ff2-dbe7e47740d1\") " Oct 11 04:25:31 crc kubenswrapper[4967]: I1011 04:25:31.323905 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd082bfe-1c3a-4801-9ff2-dbe7e47740d1-utilities" (OuterVolumeSpecName: "utilities") pod "dd082bfe-1c3a-4801-9ff2-dbe7e47740d1" (UID: "dd082bfe-1c3a-4801-9ff2-dbe7e47740d1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:25:31 crc kubenswrapper[4967]: I1011 04:25:31.330285 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/dd082bfe-1c3a-4801-9ff2-dbe7e47740d1-kube-api-access-x6q24" (OuterVolumeSpecName: "kube-api-access-x6q24") pod "dd082bfe-1c3a-4801-9ff2-dbe7e47740d1" (UID: "dd082bfe-1c3a-4801-9ff2-dbe7e47740d1"). InnerVolumeSpecName "kube-api-access-x6q24". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:25:31 crc kubenswrapper[4967]: I1011 04:25:31.377327 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/dd082bfe-1c3a-4801-9ff2-dbe7e47740d1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "dd082bfe-1c3a-4801-9ff2-dbe7e47740d1" (UID: "dd082bfe-1c3a-4801-9ff2-dbe7e47740d1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:25:31 crc kubenswrapper[4967]: I1011 04:25:31.425355 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/dd082bfe-1c3a-4801-9ff2-dbe7e47740d1-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 04:25:31 crc kubenswrapper[4967]: I1011 04:25:31.425391 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x6q24\" (UniqueName: \"kubernetes.io/projected/dd082bfe-1c3a-4801-9ff2-dbe7e47740d1-kube-api-access-x6q24\") on node \"crc\" DevicePath \"\"" Oct 11 04:25:31 crc kubenswrapper[4967]: I1011 04:25:31.425403 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/dd082bfe-1c3a-4801-9ff2-dbe7e47740d1-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 04:25:31 crc kubenswrapper[4967]: I1011 04:25:31.462987 4967 generic.go:334] "Generic (PLEG): container finished" podID="dd082bfe-1c3a-4801-9ff2-dbe7e47740d1" containerID="326cf2e8aff7bd2982619535424d0d2e23e8c8031fbf0879b0520e83419671a6" exitCode=0 Oct 11 04:25:31 crc kubenswrapper[4967]: I1011 04:25:31.463062 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k454s" event={"ID":"dd082bfe-1c3a-4801-9ff2-dbe7e47740d1","Type":"ContainerDied","Data":"326cf2e8aff7bd2982619535424d0d2e23e8c8031fbf0879b0520e83419671a6"} Oct 11 04:25:31 crc kubenswrapper[4967]: I1011 04:25:31.463125 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-k454s" event={"ID":"dd082bfe-1c3a-4801-9ff2-dbe7e47740d1","Type":"ContainerDied","Data":"344ac7c54888bc157c2c850a62d71b2c4caff85abc9c47de85322cf05d3a5fbd"} Oct 11 04:25:31 crc kubenswrapper[4967]: I1011 04:25:31.463147 4967 scope.go:117] "RemoveContainer" containerID="326cf2e8aff7bd2982619535424d0d2e23e8c8031fbf0879b0520e83419671a6" Oct 11 04:25:31 crc kubenswrapper[4967]: I1011 04:25:31.463154 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-k454s" Oct 11 04:25:31 crc kubenswrapper[4967]: I1011 04:25:31.513327 4967 scope.go:117] "RemoveContainer" containerID="52baddc35a310d149c81b0138b4df19d53374f55b9de280f7ff41ff8595f55b7" Oct 11 04:25:31 crc kubenswrapper[4967]: I1011 04:25:31.533838 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-k454s"] Oct 11 04:25:31 crc kubenswrapper[4967]: I1011 04:25:31.541559 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-k454s"] Oct 11 04:25:31 crc kubenswrapper[4967]: I1011 04:25:31.546167 4967 scope.go:117] "RemoveContainer" containerID="0edc7ef82b8f206fed69ef8bde52a262f731fc1414ad1dd6f00d861ef630a5c8" Oct 11 04:25:31 crc kubenswrapper[4967]: I1011 04:25:31.592397 4967 scope.go:117] "RemoveContainer" containerID="326cf2e8aff7bd2982619535424d0d2e23e8c8031fbf0879b0520e83419671a6" Oct 11 04:25:31 crc kubenswrapper[4967]: E1011 04:25:31.592869 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"326cf2e8aff7bd2982619535424d0d2e23e8c8031fbf0879b0520e83419671a6\": container with ID starting with 326cf2e8aff7bd2982619535424d0d2e23e8c8031fbf0879b0520e83419671a6 not found: ID does not exist" containerID="326cf2e8aff7bd2982619535424d0d2e23e8c8031fbf0879b0520e83419671a6" Oct 11 04:25:31 crc kubenswrapper[4967]: I1011 04:25:31.592902 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"326cf2e8aff7bd2982619535424d0d2e23e8c8031fbf0879b0520e83419671a6"} err="failed to get container status \"326cf2e8aff7bd2982619535424d0d2e23e8c8031fbf0879b0520e83419671a6\": rpc error: code = NotFound desc = could not find container \"326cf2e8aff7bd2982619535424d0d2e23e8c8031fbf0879b0520e83419671a6\": container with ID starting with 326cf2e8aff7bd2982619535424d0d2e23e8c8031fbf0879b0520e83419671a6 not found: ID does not exist" Oct 11 04:25:31 crc kubenswrapper[4967]: I1011 04:25:31.592923 4967 scope.go:117] "RemoveContainer" containerID="52baddc35a310d149c81b0138b4df19d53374f55b9de280f7ff41ff8595f55b7" Oct 11 04:25:31 crc kubenswrapper[4967]: E1011 04:25:31.593367 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"52baddc35a310d149c81b0138b4df19d53374f55b9de280f7ff41ff8595f55b7\": container with ID starting with 52baddc35a310d149c81b0138b4df19d53374f55b9de280f7ff41ff8595f55b7 not found: ID does not exist" containerID="52baddc35a310d149c81b0138b4df19d53374f55b9de280f7ff41ff8595f55b7" Oct 11 04:25:31 crc kubenswrapper[4967]: I1011 04:25:31.593408 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"52baddc35a310d149c81b0138b4df19d53374f55b9de280f7ff41ff8595f55b7"} err="failed to get container status \"52baddc35a310d149c81b0138b4df19d53374f55b9de280f7ff41ff8595f55b7\": rpc error: code = NotFound desc = could not find container \"52baddc35a310d149c81b0138b4df19d53374f55b9de280f7ff41ff8595f55b7\": container with ID starting with 52baddc35a310d149c81b0138b4df19d53374f55b9de280f7ff41ff8595f55b7 not found: ID does not exist" Oct 11 04:25:31 crc kubenswrapper[4967]: I1011 04:25:31.593436 4967 scope.go:117] "RemoveContainer" containerID="0edc7ef82b8f206fed69ef8bde52a262f731fc1414ad1dd6f00d861ef630a5c8" Oct 11 04:25:31 crc kubenswrapper[4967]: E1011 04:25:31.593791 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"0edc7ef82b8f206fed69ef8bde52a262f731fc1414ad1dd6f00d861ef630a5c8\": container with ID starting with 0edc7ef82b8f206fed69ef8bde52a262f731fc1414ad1dd6f00d861ef630a5c8 not found: ID does not exist" containerID="0edc7ef82b8f206fed69ef8bde52a262f731fc1414ad1dd6f00d861ef630a5c8" Oct 11 04:25:31 crc kubenswrapper[4967]: I1011 04:25:31.593825 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"0edc7ef82b8f206fed69ef8bde52a262f731fc1414ad1dd6f00d861ef630a5c8"} err="failed to get container status \"0edc7ef82b8f206fed69ef8bde52a262f731fc1414ad1dd6f00d861ef630a5c8\": rpc error: code = NotFound desc = could not find container \"0edc7ef82b8f206fed69ef8bde52a262f731fc1414ad1dd6f00d861ef630a5c8\": container with ID starting with 0edc7ef82b8f206fed69ef8bde52a262f731fc1414ad1dd6f00d861ef630a5c8 not found: ID does not exist" Oct 11 04:25:32 crc kubenswrapper[4967]: I1011 04:25:32.828764 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="dd082bfe-1c3a-4801-9ff2-dbe7e47740d1" path="/var/lib/kubelet/pods/dd082bfe-1c3a-4801-9ff2-dbe7e47740d1/volumes" Oct 11 04:25:38 crc kubenswrapper[4967]: I1011 04:25:38.574241 4967 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-h29dw" Oct 11 04:25:38 crc kubenswrapper[4967]: I1011 04:25:38.630946 4967 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-h29dw" Oct 11 04:25:38 crc kubenswrapper[4967]: I1011 04:25:38.811819 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h29dw"] Oct 11 04:25:40 crc kubenswrapper[4967]: I1011 04:25:40.111545 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-5b446d88c5-ml5f9_be6fdb4b-abb9-4e76-be7e-0c7c579d3bf8/cert-manager-controller/0.log" Oct 11 04:25:40 crc kubenswrapper[4967]: I1011 04:25:40.260410 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-7f985d654d-ncqxv_822cb861-f537-4669-8fc9-168f052a4b5d/cert-manager-cainjector/0.log" Oct 11 04:25:40 crc kubenswrapper[4967]: I1011 04:25:40.305703 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-5655c58dd6-xmk8c_ac43a6d2-5018-4d03-9a26-5da4f2619403/cert-manager-webhook/0.log" Oct 11 04:25:40 crc kubenswrapper[4967]: I1011 04:25:40.545101 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-h29dw" podUID="343b1a7b-fa49-4f53-ae16-2a827ed81936" containerName="registry-server" containerID="cri-o://94e5bfb7ca7b0e033e3d140b2e30e2cc85b331459f8f73bb61685d098542457b" gracePeriod=2 Oct 11 04:25:41 crc kubenswrapper[4967]: I1011 04:25:41.062648 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h29dw" Oct 11 04:25:41 crc kubenswrapper[4967]: I1011 04:25:41.111220 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-z6r2p\" (UniqueName: \"kubernetes.io/projected/343b1a7b-fa49-4f53-ae16-2a827ed81936-kube-api-access-z6r2p\") pod \"343b1a7b-fa49-4f53-ae16-2a827ed81936\" (UID: \"343b1a7b-fa49-4f53-ae16-2a827ed81936\") " Oct 11 04:25:41 crc kubenswrapper[4967]: I1011 04:25:41.111377 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/343b1a7b-fa49-4f53-ae16-2a827ed81936-catalog-content\") pod \"343b1a7b-fa49-4f53-ae16-2a827ed81936\" (UID: \"343b1a7b-fa49-4f53-ae16-2a827ed81936\") " Oct 11 04:25:41 crc kubenswrapper[4967]: I1011 04:25:41.111491 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/343b1a7b-fa49-4f53-ae16-2a827ed81936-utilities\") pod \"343b1a7b-fa49-4f53-ae16-2a827ed81936\" (UID: \"343b1a7b-fa49-4f53-ae16-2a827ed81936\") " Oct 11 04:25:41 crc kubenswrapper[4967]: I1011 04:25:41.112275 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/343b1a7b-fa49-4f53-ae16-2a827ed81936-utilities" (OuterVolumeSpecName: "utilities") pod "343b1a7b-fa49-4f53-ae16-2a827ed81936" (UID: "343b1a7b-fa49-4f53-ae16-2a827ed81936"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:25:41 crc kubenswrapper[4967]: I1011 04:25:41.128223 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/343b1a7b-fa49-4f53-ae16-2a827ed81936-kube-api-access-z6r2p" (OuterVolumeSpecName: "kube-api-access-z6r2p") pod "343b1a7b-fa49-4f53-ae16-2a827ed81936" (UID: "343b1a7b-fa49-4f53-ae16-2a827ed81936"). InnerVolumeSpecName "kube-api-access-z6r2p". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:25:41 crc kubenswrapper[4967]: I1011 04:25:41.203780 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/343b1a7b-fa49-4f53-ae16-2a827ed81936-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "343b1a7b-fa49-4f53-ae16-2a827ed81936" (UID: "343b1a7b-fa49-4f53-ae16-2a827ed81936"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:25:41 crc kubenswrapper[4967]: I1011 04:25:41.214623 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-z6r2p\" (UniqueName: \"kubernetes.io/projected/343b1a7b-fa49-4f53-ae16-2a827ed81936-kube-api-access-z6r2p\") on node \"crc\" DevicePath \"\"" Oct 11 04:25:41 crc kubenswrapper[4967]: I1011 04:25:41.214655 4967 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/343b1a7b-fa49-4f53-ae16-2a827ed81936-catalog-content\") on node \"crc\" DevicePath \"\"" Oct 11 04:25:41 crc kubenswrapper[4967]: I1011 04:25:41.214688 4967 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/343b1a7b-fa49-4f53-ae16-2a827ed81936-utilities\") on node \"crc\" DevicePath \"\"" Oct 11 04:25:41 crc kubenswrapper[4967]: I1011 04:25:41.558665 4967 generic.go:334] "Generic (PLEG): container finished" podID="343b1a7b-fa49-4f53-ae16-2a827ed81936" containerID="94e5bfb7ca7b0e033e3d140b2e30e2cc85b331459f8f73bb61685d098542457b" exitCode=0 Oct 11 04:25:41 crc kubenswrapper[4967]: I1011 04:25:41.558708 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h29dw" event={"ID":"343b1a7b-fa49-4f53-ae16-2a827ed81936","Type":"ContainerDied","Data":"94e5bfb7ca7b0e033e3d140b2e30e2cc85b331459f8f73bb61685d098542457b"} Oct 11 04:25:41 crc kubenswrapper[4967]: I1011 04:25:41.558733 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-h29dw" event={"ID":"343b1a7b-fa49-4f53-ae16-2a827ed81936","Type":"ContainerDied","Data":"72292d36e85690940039f619eba12d6b108183f7bffb28ede608a6faa10f6250"} Oct 11 04:25:41 crc kubenswrapper[4967]: I1011 04:25:41.558733 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-h29dw" Oct 11 04:25:41 crc kubenswrapper[4967]: I1011 04:25:41.558766 4967 scope.go:117] "RemoveContainer" containerID="94e5bfb7ca7b0e033e3d140b2e30e2cc85b331459f8f73bb61685d098542457b" Oct 11 04:25:41 crc kubenswrapper[4967]: I1011 04:25:41.589541 4967 scope.go:117] "RemoveContainer" containerID="1da7b05790ce1a95ae995901cc3ba6ad3dace5e6f1468b56fd12ffd92139fbbe" Oct 11 04:25:41 crc kubenswrapper[4967]: I1011 04:25:41.591932 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-h29dw"] Oct 11 04:25:41 crc kubenswrapper[4967]: I1011 04:25:41.603028 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-h29dw"] Oct 11 04:25:41 crc kubenswrapper[4967]: I1011 04:25:41.616163 4967 scope.go:117] "RemoveContainer" containerID="f95d1f5d07e67406dcf99f3c920a25533d38388bc349356588e1b6790ec4b68e" Oct 11 04:25:41 crc kubenswrapper[4967]: I1011 04:25:41.658317 4967 scope.go:117] "RemoveContainer" containerID="94e5bfb7ca7b0e033e3d140b2e30e2cc85b331459f8f73bb61685d098542457b" Oct 11 04:25:41 crc kubenswrapper[4967]: E1011 04:25:41.659592 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"94e5bfb7ca7b0e033e3d140b2e30e2cc85b331459f8f73bb61685d098542457b\": container with ID starting with 94e5bfb7ca7b0e033e3d140b2e30e2cc85b331459f8f73bb61685d098542457b not found: ID does not exist" containerID="94e5bfb7ca7b0e033e3d140b2e30e2cc85b331459f8f73bb61685d098542457b" Oct 11 04:25:41 crc kubenswrapper[4967]: I1011 04:25:41.659640 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"94e5bfb7ca7b0e033e3d140b2e30e2cc85b331459f8f73bb61685d098542457b"} err="failed to get container status \"94e5bfb7ca7b0e033e3d140b2e30e2cc85b331459f8f73bb61685d098542457b\": rpc error: code = NotFound desc = could not find container \"94e5bfb7ca7b0e033e3d140b2e30e2cc85b331459f8f73bb61685d098542457b\": container with ID starting with 94e5bfb7ca7b0e033e3d140b2e30e2cc85b331459f8f73bb61685d098542457b not found: ID does not exist" Oct 11 04:25:41 crc kubenswrapper[4967]: I1011 04:25:41.659668 4967 scope.go:117] "RemoveContainer" containerID="1da7b05790ce1a95ae995901cc3ba6ad3dace5e6f1468b56fd12ffd92139fbbe" Oct 11 04:25:41 crc kubenswrapper[4967]: E1011 04:25:41.660203 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1da7b05790ce1a95ae995901cc3ba6ad3dace5e6f1468b56fd12ffd92139fbbe\": container with ID starting with 1da7b05790ce1a95ae995901cc3ba6ad3dace5e6f1468b56fd12ffd92139fbbe not found: ID does not exist" containerID="1da7b05790ce1a95ae995901cc3ba6ad3dace5e6f1468b56fd12ffd92139fbbe" Oct 11 04:25:41 crc kubenswrapper[4967]: I1011 04:25:41.660264 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1da7b05790ce1a95ae995901cc3ba6ad3dace5e6f1468b56fd12ffd92139fbbe"} err="failed to get container status \"1da7b05790ce1a95ae995901cc3ba6ad3dace5e6f1468b56fd12ffd92139fbbe\": rpc error: code = NotFound desc = could not find container \"1da7b05790ce1a95ae995901cc3ba6ad3dace5e6f1468b56fd12ffd92139fbbe\": container with ID starting with 1da7b05790ce1a95ae995901cc3ba6ad3dace5e6f1468b56fd12ffd92139fbbe not found: ID does not exist" Oct 11 04:25:41 crc kubenswrapper[4967]: I1011 04:25:41.660306 4967 scope.go:117] "RemoveContainer" containerID="f95d1f5d07e67406dcf99f3c920a25533d38388bc349356588e1b6790ec4b68e" Oct 11 04:25:41 crc kubenswrapper[4967]: E1011 04:25:41.660665 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f95d1f5d07e67406dcf99f3c920a25533d38388bc349356588e1b6790ec4b68e\": container with ID starting with f95d1f5d07e67406dcf99f3c920a25533d38388bc349356588e1b6790ec4b68e not found: ID does not exist" containerID="f95d1f5d07e67406dcf99f3c920a25533d38388bc349356588e1b6790ec4b68e" Oct 11 04:25:41 crc kubenswrapper[4967]: I1011 04:25:41.660706 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f95d1f5d07e67406dcf99f3c920a25533d38388bc349356588e1b6790ec4b68e"} err="failed to get container status \"f95d1f5d07e67406dcf99f3c920a25533d38388bc349356588e1b6790ec4b68e\": rpc error: code = NotFound desc = could not find container \"f95d1f5d07e67406dcf99f3c920a25533d38388bc349356588e1b6790ec4b68e\": container with ID starting with f95d1f5d07e67406dcf99f3c920a25533d38388bc349356588e1b6790ec4b68e not found: ID does not exist" Oct 11 04:25:42 crc kubenswrapper[4967]: I1011 04:25:42.833858 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="343b1a7b-fa49-4f53-ae16-2a827ed81936" path="/var/lib/kubelet/pods/343b1a7b-fa49-4f53-ae16-2a827ed81936/volumes" Oct 11 04:25:52 crc kubenswrapper[4967]: I1011 04:25:52.793141 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-console-plugin-6b874cbd85-kfkw5_215e47d4-e289-4a93-91a7-f64b901bebeb/nmstate-console-plugin/0.log" Oct 11 04:25:52 crc kubenswrapper[4967]: I1011 04:25:52.946403 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-handler-6qvqt_b5c62540-5e0d-4601-9d0b-bef391800fe7/nmstate-handler/0.log" Oct 11 04:25:53 crc kubenswrapper[4967]: I1011 04:25:53.025309 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-bhrz7_abab15af-313d-42ef-96ce-4c0507429f93/kube-rbac-proxy/0.log" Oct 11 04:25:53 crc kubenswrapper[4967]: I1011 04:25:53.094655 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-metrics-fdff9cb8d-bhrz7_abab15af-313d-42ef-96ce-4c0507429f93/nmstate-metrics/0.log" Oct 11 04:25:53 crc kubenswrapper[4967]: I1011 04:25:53.154732 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-operator-858ddd8f98-x6r8r_419acadb-11b8-4e90-83f8-b02a39264a04/nmstate-operator/0.log" Oct 11 04:25:53 crc kubenswrapper[4967]: I1011 04:25:53.296627 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-nmstate_nmstate-webhook-6cdbc54649-4lvvm_c23b1485-a6f5-4504-bc81-a069d2a96182/nmstate-webhook/0.log" Oct 11 04:26:07 crc kubenswrapper[4967]: I1011 04:26:07.949623 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-rv747_7fb0bb1d-4f24-4ca2-8148-ac03ee0083c5/kube-rbac-proxy/0.log" Oct 11 04:26:08 crc kubenswrapper[4967]: I1011 04:26:08.079861 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_controller-68d546b9d8-rv747_7fb0bb1d-4f24-4ca2-8148-ac03ee0083c5/controller/0.log" Oct 11 04:26:08 crc kubenswrapper[4967]: I1011 04:26:08.183679 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/cp-frr-files/0.log" Oct 11 04:26:08 crc kubenswrapper[4967]: I1011 04:26:08.339381 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/cp-metrics/0.log" Oct 11 04:26:08 crc kubenswrapper[4967]: I1011 04:26:08.363449 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/cp-reloader/0.log" Oct 11 04:26:08 crc kubenswrapper[4967]: I1011 04:26:08.378524 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/cp-reloader/0.log" Oct 11 04:26:08 crc kubenswrapper[4967]: I1011 04:26:08.386468 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/cp-frr-files/0.log" Oct 11 04:26:08 crc kubenswrapper[4967]: I1011 04:26:08.539876 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/cp-frr-files/0.log" Oct 11 04:26:08 crc kubenswrapper[4967]: I1011 04:26:08.567557 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/cp-metrics/0.log" Oct 11 04:26:08 crc kubenswrapper[4967]: I1011 04:26:08.568108 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/cp-reloader/0.log" Oct 11 04:26:08 crc kubenswrapper[4967]: I1011 04:26:08.604609 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/cp-metrics/0.log" Oct 11 04:26:08 crc kubenswrapper[4967]: I1011 04:26:08.741048 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/cp-reloader/0.log" Oct 11 04:26:08 crc kubenswrapper[4967]: I1011 04:26:08.743263 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/cp-frr-files/0.log" Oct 11 04:26:08 crc kubenswrapper[4967]: I1011 04:26:08.782378 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/cp-metrics/0.log" Oct 11 04:26:08 crc kubenswrapper[4967]: I1011 04:26:08.808010 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/controller/0.log" Oct 11 04:26:08 crc kubenswrapper[4967]: I1011 04:26:08.966430 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/frr-metrics/0.log" Oct 11 04:26:09 crc kubenswrapper[4967]: I1011 04:26:09.002243 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/kube-rbac-proxy/0.log" Oct 11 04:26:09 crc kubenswrapper[4967]: I1011 04:26:09.037170 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/kube-rbac-proxy-frr/0.log" Oct 11 04:26:09 crc kubenswrapper[4967]: I1011 04:26:09.204488 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/reloader/0.log" Oct 11 04:26:09 crc kubenswrapper[4967]: I1011 04:26:09.258171 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-webhook-server-64bf5d555-j9lg6_50110bd6-5363-43c8-b3a8-b4b3762c0c43/frr-k8s-webhook-server/0.log" Oct 11 04:26:09 crc kubenswrapper[4967]: I1011 04:26:09.497332 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-controller-manager-5cfc444676-2s7hf_f3173c9b-89b4-420d-a852-f293d581da52/manager/0.log" Oct 11 04:26:09 crc kubenswrapper[4967]: I1011 04:26:09.562716 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_metallb-operator-webhook-server-56f68cd588-8fj48_c88642bc-96a7-4662-a035-c2c6043e322f/webhook-server/0.log" Oct 11 04:26:09 crc kubenswrapper[4967]: I1011 04:26:09.715389 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-np2lm_816b70d7-ff40-4cee-b21d-01008f0d4868/kube-rbac-proxy/0.log" Oct 11 04:26:09 crc kubenswrapper[4967]: I1011 04:26:09.986317 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_frr-k8s-n67vr_e9d93993-50a0-468a-a3db-134db9acd6ba/frr/0.log" Oct 11 04:26:10 crc kubenswrapper[4967]: I1011 04:26:10.099857 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/metallb-system_speaker-np2lm_816b70d7-ff40-4cee-b21d-01008f0d4868/speaker/0.log" Oct 11 04:26:22 crc kubenswrapper[4967]: I1011 04:26:22.196540 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn_bd662d9c-6ab0-4518-8fab-83115ccd3569/util/0.log" Oct 11 04:26:22 crc kubenswrapper[4967]: I1011 04:26:22.324109 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn_bd662d9c-6ab0-4518-8fab-83115ccd3569/util/0.log" Oct 11 04:26:22 crc kubenswrapper[4967]: I1011 04:26:22.397362 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn_bd662d9c-6ab0-4518-8fab-83115ccd3569/pull/0.log" Oct 11 04:26:22 crc kubenswrapper[4967]: I1011 04:26:22.423790 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn_bd662d9c-6ab0-4518-8fab-83115ccd3569/pull/0.log" Oct 11 04:26:22 crc kubenswrapper[4967]: I1011 04:26:22.564521 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn_bd662d9c-6ab0-4518-8fab-83115ccd3569/pull/0.log" Oct 11 04:26:22 crc kubenswrapper[4967]: I1011 04:26:22.581531 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn_bd662d9c-6ab0-4518-8fab-83115ccd3569/extract/0.log" Oct 11 04:26:22 crc kubenswrapper[4967]: I1011 04:26:22.596138 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_8f2f4ee801e5826a37d84a7b1fc4ccbf6b79de668302737d0f1152d8d2czgkn_bd662d9c-6ab0-4518-8fab-83115ccd3569/util/0.log" Oct 11 04:26:22 crc kubenswrapper[4967]: I1011 04:26:22.776841 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mxkbg_afb60be7-f83c-4dc3-9129-ec66a8b75a55/extract-utilities/0.log" Oct 11 04:26:22 crc kubenswrapper[4967]: I1011 04:26:22.951012 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mxkbg_afb60be7-f83c-4dc3-9129-ec66a8b75a55/extract-content/0.log" Oct 11 04:26:22 crc kubenswrapper[4967]: I1011 04:26:22.952642 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mxkbg_afb60be7-f83c-4dc3-9129-ec66a8b75a55/extract-content/0.log" Oct 11 04:26:22 crc kubenswrapper[4967]: I1011 04:26:22.955548 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mxkbg_afb60be7-f83c-4dc3-9129-ec66a8b75a55/extract-utilities/0.log" Oct 11 04:26:23 crc kubenswrapper[4967]: I1011 04:26:23.112031 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mxkbg_afb60be7-f83c-4dc3-9129-ec66a8b75a55/extract-utilities/0.log" Oct 11 04:26:23 crc kubenswrapper[4967]: I1011 04:26:23.147043 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mxkbg_afb60be7-f83c-4dc3-9129-ec66a8b75a55/extract-content/0.log" Oct 11 04:26:23 crc kubenswrapper[4967]: I1011 04:26:23.318859 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dgst6_cd79944d-93c5-45d4-b68d-5cff92c3d111/extract-utilities/0.log" Oct 11 04:26:23 crc kubenswrapper[4967]: I1011 04:26:23.373713 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-mxkbg_afb60be7-f83c-4dc3-9129-ec66a8b75a55/registry-server/0.log" Oct 11 04:26:23 crc kubenswrapper[4967]: I1011 04:26:23.511532 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dgst6_cd79944d-93c5-45d4-b68d-5cff92c3d111/extract-utilities/0.log" Oct 11 04:26:23 crc kubenswrapper[4967]: I1011 04:26:23.518222 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dgst6_cd79944d-93c5-45d4-b68d-5cff92c3d111/extract-content/0.log" Oct 11 04:26:23 crc kubenswrapper[4967]: I1011 04:26:23.550061 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dgst6_cd79944d-93c5-45d4-b68d-5cff92c3d111/extract-content/0.log" Oct 11 04:26:23 crc kubenswrapper[4967]: I1011 04:26:23.700204 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dgst6_cd79944d-93c5-45d4-b68d-5cff92c3d111/extract-content/0.log" Oct 11 04:26:23 crc kubenswrapper[4967]: I1011 04:26:23.712726 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dgst6_cd79944d-93c5-45d4-b68d-5cff92c3d111/extract-utilities/0.log" Oct 11 04:26:23 crc kubenswrapper[4967]: I1011 04:26:23.901762 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6_afec4044-82ab-4b5b-8669-7605ab680e10/util/0.log" Oct 11 04:26:24 crc kubenswrapper[4967]: I1011 04:26:24.121269 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dgst6_cd79944d-93c5-45d4-b68d-5cff92c3d111/registry-server/0.log" Oct 11 04:26:24 crc kubenswrapper[4967]: I1011 04:26:24.127774 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6_afec4044-82ab-4b5b-8669-7605ab680e10/util/0.log" Oct 11 04:26:24 crc kubenswrapper[4967]: I1011 04:26:24.130840 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6_afec4044-82ab-4b5b-8669-7605ab680e10/pull/0.log" Oct 11 04:26:24 crc kubenswrapper[4967]: I1011 04:26:24.145958 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6_afec4044-82ab-4b5b-8669-7605ab680e10/pull/0.log" Oct 11 04:26:24 crc kubenswrapper[4967]: I1011 04:26:24.332410 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6_afec4044-82ab-4b5b-8669-7605ab680e10/extract/0.log" Oct 11 04:26:24 crc kubenswrapper[4967]: I1011 04:26:24.336546 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6_afec4044-82ab-4b5b-8669-7605ab680e10/util/0.log" Oct 11 04:26:24 crc kubenswrapper[4967]: I1011 04:26:24.354867 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_fa9831ede5d93c33d525b70ce6ddf94e500d80992af75a3305fe98835chdjn6_afec4044-82ab-4b5b-8669-7605ab680e10/pull/0.log" Oct 11 04:26:24 crc kubenswrapper[4967]: I1011 04:26:24.502100 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-sp7h8_020dc4f8-c49b-4447-8373-426bf747cace/marketplace-operator/0.log" Oct 11 04:26:24 crc kubenswrapper[4967]: I1011 04:26:24.533435 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p2xpx_065a2523-459f-440e-b5da-5ea5318fd89f/extract-utilities/0.log" Oct 11 04:26:24 crc kubenswrapper[4967]: I1011 04:26:24.715709 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p2xpx_065a2523-459f-440e-b5da-5ea5318fd89f/extract-content/0.log" Oct 11 04:26:24 crc kubenswrapper[4967]: I1011 04:26:24.720550 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p2xpx_065a2523-459f-440e-b5da-5ea5318fd89f/extract-utilities/0.log" Oct 11 04:26:24 crc kubenswrapper[4967]: I1011 04:26:24.736175 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p2xpx_065a2523-459f-440e-b5da-5ea5318fd89f/extract-content/0.log" Oct 11 04:26:24 crc kubenswrapper[4967]: I1011 04:26:24.899191 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p2xpx_065a2523-459f-440e-b5da-5ea5318fd89f/extract-utilities/0.log" Oct 11 04:26:24 crc kubenswrapper[4967]: I1011 04:26:24.911105 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p2xpx_065a2523-459f-440e-b5da-5ea5318fd89f/extract-content/0.log" Oct 11 04:26:25 crc kubenswrapper[4967]: I1011 04:26:25.025032 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-marketplace-p2xpx_065a2523-459f-440e-b5da-5ea5318fd89f/registry-server/0.log" Oct 11 04:26:25 crc kubenswrapper[4967]: I1011 04:26:25.077101 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-khtzc_553fcde4-86f4-4a03-9650-bb6bd3ee5d0e/extract-utilities/0.log" Oct 11 04:26:25 crc kubenswrapper[4967]: I1011 04:26:25.202107 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-khtzc_553fcde4-86f4-4a03-9650-bb6bd3ee5d0e/extract-utilities/0.log" Oct 11 04:26:25 crc kubenswrapper[4967]: I1011 04:26:25.245204 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-khtzc_553fcde4-86f4-4a03-9650-bb6bd3ee5d0e/extract-content/0.log" Oct 11 04:26:25 crc kubenswrapper[4967]: I1011 04:26:25.251526 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-khtzc_553fcde4-86f4-4a03-9650-bb6bd3ee5d0e/extract-content/0.log" Oct 11 04:26:25 crc kubenswrapper[4967]: I1011 04:26:25.412058 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-khtzc_553fcde4-86f4-4a03-9650-bb6bd3ee5d0e/extract-utilities/0.log" Oct 11 04:26:25 crc kubenswrapper[4967]: I1011 04:26:25.416370 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-khtzc_553fcde4-86f4-4a03-9650-bb6bd3ee5d0e/extract-content/0.log" Oct 11 04:26:25 crc kubenswrapper[4967]: I1011 04:26:25.752858 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-khtzc_553fcde4-86f4-4a03-9650-bb6bd3ee5d0e/registry-server/0.log" Oct 11 04:26:42 crc kubenswrapper[4967]: E1011 04:26:42.637602 4967 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.50:42010->38.102.83.50:40947: write tcp 38.102.83.50:42010->38.102.83.50:40947: write: broken pipe Oct 11 04:26:44 crc kubenswrapper[4967]: I1011 04:26:44.095774 4967 patch_prober.go:28] interesting pod/machine-config-daemon-sgrv2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:26:44 crc kubenswrapper[4967]: I1011 04:26:44.096130 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:26:57 crc kubenswrapper[4967]: E1011 04:26:57.938705 4967 upgradeaware.go:427] Error proxying data from client to backend: readfrom tcp 38.102.83.50:34928->38.102.83.50:40947: write tcp 38.102.83.50:34928->38.102.83.50:40947: write: connection reset by peer Oct 11 04:27:14 crc kubenswrapper[4967]: I1011 04:27:14.088513 4967 patch_prober.go:28] interesting pod/machine-config-daemon-sgrv2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:27:14 crc kubenswrapper[4967]: I1011 04:27:14.089058 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:27:44 crc kubenswrapper[4967]: I1011 04:27:44.088614 4967 patch_prober.go:28] interesting pod/machine-config-daemon-sgrv2 container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Oct 11 04:27:44 crc kubenswrapper[4967]: I1011 04:27:44.090380 4967 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Oct 11 04:27:44 crc kubenswrapper[4967]: I1011 04:27:44.090568 4967 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" Oct 11 04:27:44 crc kubenswrapper[4967]: I1011 04:27:44.091417 4967 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"dfd2cbbb285a882d3d3b0918a1f4268e32c2905a0b9d3bb49ba405a42cb0df28"} pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Oct 11 04:27:44 crc kubenswrapper[4967]: I1011 04:27:44.091571 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" containerName="machine-config-daemon" containerID="cri-o://dfd2cbbb285a882d3d3b0918a1f4268e32c2905a0b9d3bb49ba405a42cb0df28" gracePeriod=600 Oct 11 04:27:44 crc kubenswrapper[4967]: E1011 04:27:44.259518 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:27:44 crc kubenswrapper[4967]: I1011 04:27:44.808901 4967 generic.go:334] "Generic (PLEG): container finished" podID="97c07678-14be-410c-b61f-498cb49bc960" containerID="dfd2cbbb285a882d3d3b0918a1f4268e32c2905a0b9d3bb49ba405a42cb0df28" exitCode=0 Oct 11 04:27:44 crc kubenswrapper[4967]: I1011 04:27:44.808980 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" event={"ID":"97c07678-14be-410c-b61f-498cb49bc960","Type":"ContainerDied","Data":"dfd2cbbb285a882d3d3b0918a1f4268e32c2905a0b9d3bb49ba405a42cb0df28"} Oct 11 04:27:44 crc kubenswrapper[4967]: I1011 04:27:44.809037 4967 scope.go:117] "RemoveContainer" containerID="d6d902b7233db28561ab70d9cf62192d380aef02bd4e78fc79f31d3c320df63c" Oct 11 04:27:44 crc kubenswrapper[4967]: I1011 04:27:44.823276 4967 scope.go:117] "RemoveContainer" containerID="dfd2cbbb285a882d3d3b0918a1f4268e32c2905a0b9d3bb49ba405a42cb0df28" Oct 11 04:27:44 crc kubenswrapper[4967]: E1011 04:27:44.823949 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:27:51 crc kubenswrapper[4967]: I1011 04:27:51.911409 4967 generic.go:334] "Generic (PLEG): container finished" podID="393350ef-d545-42cd-92d2-59ee992dd2bc" containerID="5713d3fe8a8383721a12fa412cb095b6b308a0fd981cc6327434293fca4a7482" exitCode=0 Oct 11 04:27:51 crc kubenswrapper[4967]: I1011 04:27:51.911516 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-rtxqh/must-gather-4zlrw" event={"ID":"393350ef-d545-42cd-92d2-59ee992dd2bc","Type":"ContainerDied","Data":"5713d3fe8a8383721a12fa412cb095b6b308a0fd981cc6327434293fca4a7482"} Oct 11 04:27:51 crc kubenswrapper[4967]: I1011 04:27:51.913019 4967 scope.go:117] "RemoveContainer" containerID="5713d3fe8a8383721a12fa412cb095b6b308a0fd981cc6327434293fca4a7482" Oct 11 04:27:52 crc kubenswrapper[4967]: I1011 04:27:52.108808 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-rtxqh_must-gather-4zlrw_393350ef-d545-42cd-92d2-59ee992dd2bc/gather/0.log" Oct 11 04:27:56 crc kubenswrapper[4967]: I1011 04:27:56.814985 4967 scope.go:117] "RemoveContainer" containerID="dfd2cbbb285a882d3d3b0918a1f4268e32c2905a0b9d3bb49ba405a42cb0df28" Oct 11 04:27:56 crc kubenswrapper[4967]: E1011 04:27:56.816238 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:28:00 crc kubenswrapper[4967]: I1011 04:28:00.938520 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-rtxqh/must-gather-4zlrw"] Oct 11 04:28:00 crc kubenswrapper[4967]: I1011 04:28:00.939396 4967 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-rtxqh/must-gather-4zlrw" podUID="393350ef-d545-42cd-92d2-59ee992dd2bc" containerName="copy" containerID="cri-o://d53cafbbae23954dcd74b593c51989cbf17bc3f0b9eb0a7708abd835965f181a" gracePeriod=2 Oct 11 04:28:00 crc kubenswrapper[4967]: I1011 04:28:00.951403 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-rtxqh/must-gather-4zlrw"] Oct 11 04:28:01 crc kubenswrapper[4967]: I1011 04:28:01.462263 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-rtxqh_must-gather-4zlrw_393350ef-d545-42cd-92d2-59ee992dd2bc/copy/0.log" Oct 11 04:28:01 crc kubenswrapper[4967]: I1011 04:28:01.462900 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rtxqh/must-gather-4zlrw" Oct 11 04:28:01 crc kubenswrapper[4967]: I1011 04:28:01.650397 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/393350ef-d545-42cd-92d2-59ee992dd2bc-must-gather-output\") pod \"393350ef-d545-42cd-92d2-59ee992dd2bc\" (UID: \"393350ef-d545-42cd-92d2-59ee992dd2bc\") " Oct 11 04:28:01 crc kubenswrapper[4967]: I1011 04:28:01.650493 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-q58xk\" (UniqueName: \"kubernetes.io/projected/393350ef-d545-42cd-92d2-59ee992dd2bc-kube-api-access-q58xk\") pod \"393350ef-d545-42cd-92d2-59ee992dd2bc\" (UID: \"393350ef-d545-42cd-92d2-59ee992dd2bc\") " Oct 11 04:28:01 crc kubenswrapper[4967]: I1011 04:28:01.670370 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/393350ef-d545-42cd-92d2-59ee992dd2bc-kube-api-access-q58xk" (OuterVolumeSpecName: "kube-api-access-q58xk") pod "393350ef-d545-42cd-92d2-59ee992dd2bc" (UID: "393350ef-d545-42cd-92d2-59ee992dd2bc"). InnerVolumeSpecName "kube-api-access-q58xk". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:28:01 crc kubenswrapper[4967]: I1011 04:28:01.754536 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-q58xk\" (UniqueName: \"kubernetes.io/projected/393350ef-d545-42cd-92d2-59ee992dd2bc-kube-api-access-q58xk\") on node \"crc\" DevicePath \"\"" Oct 11 04:28:01 crc kubenswrapper[4967]: I1011 04:28:01.794133 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/393350ef-d545-42cd-92d2-59ee992dd2bc-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "393350ef-d545-42cd-92d2-59ee992dd2bc" (UID: "393350ef-d545-42cd-92d2-59ee992dd2bc"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Oct 11 04:28:01 crc kubenswrapper[4967]: I1011 04:28:01.856787 4967 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/393350ef-d545-42cd-92d2-59ee992dd2bc-must-gather-output\") on node \"crc\" DevicePath \"\"" Oct 11 04:28:02 crc kubenswrapper[4967]: I1011 04:28:02.029563 4967 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-rtxqh_must-gather-4zlrw_393350ef-d545-42cd-92d2-59ee992dd2bc/copy/0.log" Oct 11 04:28:02 crc kubenswrapper[4967]: I1011 04:28:02.030324 4967 generic.go:334] "Generic (PLEG): container finished" podID="393350ef-d545-42cd-92d2-59ee992dd2bc" containerID="d53cafbbae23954dcd74b593c51989cbf17bc3f0b9eb0a7708abd835965f181a" exitCode=143 Oct 11 04:28:02 crc kubenswrapper[4967]: I1011 04:28:02.030366 4967 scope.go:117] "RemoveContainer" containerID="d53cafbbae23954dcd74b593c51989cbf17bc3f0b9eb0a7708abd835965f181a" Oct 11 04:28:02 crc kubenswrapper[4967]: I1011 04:28:02.030472 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-rtxqh/must-gather-4zlrw" Oct 11 04:28:02 crc kubenswrapper[4967]: I1011 04:28:02.067061 4967 scope.go:117] "RemoveContainer" containerID="5713d3fe8a8383721a12fa412cb095b6b308a0fd981cc6327434293fca4a7482" Oct 11 04:28:02 crc kubenswrapper[4967]: I1011 04:28:02.140472 4967 scope.go:117] "RemoveContainer" containerID="d53cafbbae23954dcd74b593c51989cbf17bc3f0b9eb0a7708abd835965f181a" Oct 11 04:28:02 crc kubenswrapper[4967]: E1011 04:28:02.140911 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d53cafbbae23954dcd74b593c51989cbf17bc3f0b9eb0a7708abd835965f181a\": container with ID starting with d53cafbbae23954dcd74b593c51989cbf17bc3f0b9eb0a7708abd835965f181a not found: ID does not exist" containerID="d53cafbbae23954dcd74b593c51989cbf17bc3f0b9eb0a7708abd835965f181a" Oct 11 04:28:02 crc kubenswrapper[4967]: I1011 04:28:02.140946 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d53cafbbae23954dcd74b593c51989cbf17bc3f0b9eb0a7708abd835965f181a"} err="failed to get container status \"d53cafbbae23954dcd74b593c51989cbf17bc3f0b9eb0a7708abd835965f181a\": rpc error: code = NotFound desc = could not find container \"d53cafbbae23954dcd74b593c51989cbf17bc3f0b9eb0a7708abd835965f181a\": container with ID starting with d53cafbbae23954dcd74b593c51989cbf17bc3f0b9eb0a7708abd835965f181a not found: ID does not exist" Oct 11 04:28:02 crc kubenswrapper[4967]: I1011 04:28:02.140968 4967 scope.go:117] "RemoveContainer" containerID="5713d3fe8a8383721a12fa412cb095b6b308a0fd981cc6327434293fca4a7482" Oct 11 04:28:02 crc kubenswrapper[4967]: E1011 04:28:02.141342 4967 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5713d3fe8a8383721a12fa412cb095b6b308a0fd981cc6327434293fca4a7482\": container with ID starting with 5713d3fe8a8383721a12fa412cb095b6b308a0fd981cc6327434293fca4a7482 not found: ID does not exist" containerID="5713d3fe8a8383721a12fa412cb095b6b308a0fd981cc6327434293fca4a7482" Oct 11 04:28:02 crc kubenswrapper[4967]: I1011 04:28:02.141370 4967 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5713d3fe8a8383721a12fa412cb095b6b308a0fd981cc6327434293fca4a7482"} err="failed to get container status \"5713d3fe8a8383721a12fa412cb095b6b308a0fd981cc6327434293fca4a7482\": rpc error: code = NotFound desc = could not find container \"5713d3fe8a8383721a12fa412cb095b6b308a0fd981cc6327434293fca4a7482\": container with ID starting with 5713d3fe8a8383721a12fa412cb095b6b308a0fd981cc6327434293fca4a7482 not found: ID does not exist" Oct 11 04:28:02 crc kubenswrapper[4967]: I1011 04:28:02.831627 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="393350ef-d545-42cd-92d2-59ee992dd2bc" path="/var/lib/kubelet/pods/393350ef-d545-42cd-92d2-59ee992dd2bc/volumes" Oct 11 04:28:08 crc kubenswrapper[4967]: I1011 04:28:08.815606 4967 scope.go:117] "RemoveContainer" containerID="dfd2cbbb285a882d3d3b0918a1f4268e32c2905a0b9d3bb49ba405a42cb0df28" Oct 11 04:28:08 crc kubenswrapper[4967]: E1011 04:28:08.817045 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:28:22 crc kubenswrapper[4967]: I1011 04:28:22.827169 4967 scope.go:117] "RemoveContainer" containerID="dfd2cbbb285a882d3d3b0918a1f4268e32c2905a0b9d3bb49ba405a42cb0df28" Oct 11 04:28:22 crc kubenswrapper[4967]: E1011 04:28:22.830171 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:28:37 crc kubenswrapper[4967]: I1011 04:28:37.815967 4967 scope.go:117] "RemoveContainer" containerID="dfd2cbbb285a882d3d3b0918a1f4268e32c2905a0b9d3bb49ba405a42cb0df28" Oct 11 04:28:37 crc kubenswrapper[4967]: E1011 04:28:37.817667 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:28:51 crc kubenswrapper[4967]: I1011 04:28:51.815672 4967 scope.go:117] "RemoveContainer" containerID="dfd2cbbb285a882d3d3b0918a1f4268e32c2905a0b9d3bb49ba405a42cb0df28" Oct 11 04:28:51 crc kubenswrapper[4967]: E1011 04:28:51.817063 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:29:02 crc kubenswrapper[4967]: I1011 04:29:02.851136 4967 scope.go:117] "RemoveContainer" containerID="dfd2cbbb285a882d3d3b0918a1f4268e32c2905a0b9d3bb49ba405a42cb0df28" Oct 11 04:29:02 crc kubenswrapper[4967]: E1011 04:29:02.854043 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:29:13 crc kubenswrapper[4967]: I1011 04:29:13.816047 4967 scope.go:117] "RemoveContainer" containerID="dfd2cbbb285a882d3d3b0918a1f4268e32c2905a0b9d3bb49ba405a42cb0df28" Oct 11 04:29:13 crc kubenswrapper[4967]: E1011 04:29:13.817026 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:29:27 crc kubenswrapper[4967]: I1011 04:29:27.815443 4967 scope.go:117] "RemoveContainer" containerID="dfd2cbbb285a882d3d3b0918a1f4268e32c2905a0b9d3bb49ba405a42cb0df28" Oct 11 04:29:27 crc kubenswrapper[4967]: E1011 04:29:27.816570 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:29:40 crc kubenswrapper[4967]: I1011 04:29:40.814448 4967 scope.go:117] "RemoveContainer" containerID="dfd2cbbb285a882d3d3b0918a1f4268e32c2905a0b9d3bb49ba405a42cb0df28" Oct 11 04:29:40 crc kubenswrapper[4967]: E1011 04:29:40.815269 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:29:55 crc kubenswrapper[4967]: I1011 04:29:55.814974 4967 scope.go:117] "RemoveContainer" containerID="dfd2cbbb285a882d3d3b0918a1f4268e32c2905a0b9d3bb49ba405a42cb0df28" Oct 11 04:29:55 crc kubenswrapper[4967]: E1011 04:29:55.816192 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:30:00 crc kubenswrapper[4967]: I1011 04:30:00.176677 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335950-whkkp"] Oct 11 04:30:00 crc kubenswrapper[4967]: E1011 04:30:00.177800 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="343b1a7b-fa49-4f53-ae16-2a827ed81936" containerName="registry-server" Oct 11 04:30:00 crc kubenswrapper[4967]: I1011 04:30:00.177822 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="343b1a7b-fa49-4f53-ae16-2a827ed81936" containerName="registry-server" Oct 11 04:30:00 crc kubenswrapper[4967]: E1011 04:30:00.177839 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="393350ef-d545-42cd-92d2-59ee992dd2bc" containerName="gather" Oct 11 04:30:00 crc kubenswrapper[4967]: I1011 04:30:00.177850 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="393350ef-d545-42cd-92d2-59ee992dd2bc" containerName="gather" Oct 11 04:30:00 crc kubenswrapper[4967]: E1011 04:30:00.177873 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd082bfe-1c3a-4801-9ff2-dbe7e47740d1" containerName="extract-utilities" Oct 11 04:30:00 crc kubenswrapper[4967]: I1011 04:30:00.177885 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd082bfe-1c3a-4801-9ff2-dbe7e47740d1" containerName="extract-utilities" Oct 11 04:30:00 crc kubenswrapper[4967]: E1011 04:30:00.177902 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd082bfe-1c3a-4801-9ff2-dbe7e47740d1" containerName="extract-content" Oct 11 04:30:00 crc kubenswrapper[4967]: I1011 04:30:00.177912 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd082bfe-1c3a-4801-9ff2-dbe7e47740d1" containerName="extract-content" Oct 11 04:30:00 crc kubenswrapper[4967]: E1011 04:30:00.177930 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="393350ef-d545-42cd-92d2-59ee992dd2bc" containerName="copy" Oct 11 04:30:00 crc kubenswrapper[4967]: I1011 04:30:00.177941 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="393350ef-d545-42cd-92d2-59ee992dd2bc" containerName="copy" Oct 11 04:30:00 crc kubenswrapper[4967]: E1011 04:30:00.177969 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="343b1a7b-fa49-4f53-ae16-2a827ed81936" containerName="extract-utilities" Oct 11 04:30:00 crc kubenswrapper[4967]: I1011 04:30:00.177980 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="343b1a7b-fa49-4f53-ae16-2a827ed81936" containerName="extract-utilities" Oct 11 04:30:00 crc kubenswrapper[4967]: E1011 04:30:00.178005 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8be18f4f-56aa-4c11-9870-5d9e652338cd" containerName="extract-content" Oct 11 04:30:00 crc kubenswrapper[4967]: I1011 04:30:00.178015 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="8be18f4f-56aa-4c11-9870-5d9e652338cd" containerName="extract-content" Oct 11 04:30:00 crc kubenswrapper[4967]: E1011 04:30:00.178036 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="dd082bfe-1c3a-4801-9ff2-dbe7e47740d1" containerName="registry-server" Oct 11 04:30:00 crc kubenswrapper[4967]: I1011 04:30:00.178048 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="dd082bfe-1c3a-4801-9ff2-dbe7e47740d1" containerName="registry-server" Oct 11 04:30:00 crc kubenswrapper[4967]: E1011 04:30:00.178114 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8be18f4f-56aa-4c11-9870-5d9e652338cd" containerName="extract-utilities" Oct 11 04:30:00 crc kubenswrapper[4967]: I1011 04:30:00.178127 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="8be18f4f-56aa-4c11-9870-5d9e652338cd" containerName="extract-utilities" Oct 11 04:30:00 crc kubenswrapper[4967]: E1011 04:30:00.178146 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="343b1a7b-fa49-4f53-ae16-2a827ed81936" containerName="extract-content" Oct 11 04:30:00 crc kubenswrapper[4967]: I1011 04:30:00.178155 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="343b1a7b-fa49-4f53-ae16-2a827ed81936" containerName="extract-content" Oct 11 04:30:00 crc kubenswrapper[4967]: E1011 04:30:00.178173 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8be18f4f-56aa-4c11-9870-5d9e652338cd" containerName="registry-server" Oct 11 04:30:00 crc kubenswrapper[4967]: I1011 04:30:00.178184 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="8be18f4f-56aa-4c11-9870-5d9e652338cd" containerName="registry-server" Oct 11 04:30:00 crc kubenswrapper[4967]: I1011 04:30:00.178467 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="393350ef-d545-42cd-92d2-59ee992dd2bc" containerName="copy" Oct 11 04:30:00 crc kubenswrapper[4967]: I1011 04:30:00.178504 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="393350ef-d545-42cd-92d2-59ee992dd2bc" containerName="gather" Oct 11 04:30:00 crc kubenswrapper[4967]: I1011 04:30:00.178523 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="343b1a7b-fa49-4f53-ae16-2a827ed81936" containerName="registry-server" Oct 11 04:30:00 crc kubenswrapper[4967]: I1011 04:30:00.178547 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="dd082bfe-1c3a-4801-9ff2-dbe7e47740d1" containerName="registry-server" Oct 11 04:30:00 crc kubenswrapper[4967]: I1011 04:30:00.178564 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="8be18f4f-56aa-4c11-9870-5d9e652338cd" containerName="registry-server" Oct 11 04:30:00 crc kubenswrapper[4967]: I1011 04:30:00.179544 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-whkkp" Oct 11 04:30:00 crc kubenswrapper[4967]: I1011 04:30:00.182491 4967 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Oct 11 04:30:00 crc kubenswrapper[4967]: I1011 04:30:00.184051 4967 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Oct 11 04:30:00 crc kubenswrapper[4967]: I1011 04:30:00.197803 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335950-whkkp"] Oct 11 04:30:00 crc kubenswrapper[4967]: I1011 04:30:00.292111 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eef2317b-924a-44b7-9fb8-0913e229df0f-secret-volume\") pod \"collect-profiles-29335950-whkkp\" (UID: \"eef2317b-924a-44b7-9fb8-0913e229df0f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-whkkp" Oct 11 04:30:00 crc kubenswrapper[4967]: I1011 04:30:00.292350 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eef2317b-924a-44b7-9fb8-0913e229df0f-config-volume\") pod \"collect-profiles-29335950-whkkp\" (UID: \"eef2317b-924a-44b7-9fb8-0913e229df0f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-whkkp" Oct 11 04:30:00 crc kubenswrapper[4967]: I1011 04:30:00.292721 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gxd98\" (UniqueName: \"kubernetes.io/projected/eef2317b-924a-44b7-9fb8-0913e229df0f-kube-api-access-gxd98\") pod \"collect-profiles-29335950-whkkp\" (UID: \"eef2317b-924a-44b7-9fb8-0913e229df0f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-whkkp" Oct 11 04:30:00 crc kubenswrapper[4967]: I1011 04:30:00.394483 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gxd98\" (UniqueName: \"kubernetes.io/projected/eef2317b-924a-44b7-9fb8-0913e229df0f-kube-api-access-gxd98\") pod \"collect-profiles-29335950-whkkp\" (UID: \"eef2317b-924a-44b7-9fb8-0913e229df0f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-whkkp" Oct 11 04:30:00 crc kubenswrapper[4967]: I1011 04:30:00.394650 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eef2317b-924a-44b7-9fb8-0913e229df0f-secret-volume\") pod \"collect-profiles-29335950-whkkp\" (UID: \"eef2317b-924a-44b7-9fb8-0913e229df0f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-whkkp" Oct 11 04:30:00 crc kubenswrapper[4967]: I1011 04:30:00.394881 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eef2317b-924a-44b7-9fb8-0913e229df0f-config-volume\") pod \"collect-profiles-29335950-whkkp\" (UID: \"eef2317b-924a-44b7-9fb8-0913e229df0f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-whkkp" Oct 11 04:30:00 crc kubenswrapper[4967]: I1011 04:30:00.397010 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eef2317b-924a-44b7-9fb8-0913e229df0f-config-volume\") pod \"collect-profiles-29335950-whkkp\" (UID: \"eef2317b-924a-44b7-9fb8-0913e229df0f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-whkkp" Oct 11 04:30:00 crc kubenswrapper[4967]: I1011 04:30:00.408479 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eef2317b-924a-44b7-9fb8-0913e229df0f-secret-volume\") pod \"collect-profiles-29335950-whkkp\" (UID: \"eef2317b-924a-44b7-9fb8-0913e229df0f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-whkkp" Oct 11 04:30:00 crc kubenswrapper[4967]: I1011 04:30:00.417533 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gxd98\" (UniqueName: \"kubernetes.io/projected/eef2317b-924a-44b7-9fb8-0913e229df0f-kube-api-access-gxd98\") pod \"collect-profiles-29335950-whkkp\" (UID: \"eef2317b-924a-44b7-9fb8-0913e229df0f\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-whkkp" Oct 11 04:30:00 crc kubenswrapper[4967]: I1011 04:30:00.509797 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-whkkp" Oct 11 04:30:00 crc kubenswrapper[4967]: I1011 04:30:00.974562 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335950-whkkp"] Oct 11 04:30:01 crc kubenswrapper[4967]: I1011 04:30:01.385319 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-whkkp" event={"ID":"eef2317b-924a-44b7-9fb8-0913e229df0f","Type":"ContainerStarted","Data":"0aa35a0c8da4feac2d3f08e11710e03ce2ccb39337139f9edb9878e27badd9ab"} Oct 11 04:30:01 crc kubenswrapper[4967]: I1011 04:30:01.385362 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-whkkp" event={"ID":"eef2317b-924a-44b7-9fb8-0913e229df0f","Type":"ContainerStarted","Data":"6bcd41998f5b738a32fb9b108e2462d8a01d42efada042b50a8e7aacc7c91fcf"} Oct 11 04:30:01 crc kubenswrapper[4967]: I1011 04:30:01.401139 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-whkkp" podStartSLOduration=1.401123575 podStartE2EDuration="1.401123575s" podCreationTimestamp="2025-10-11 04:30:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2025-10-11 04:30:01.40090843 +0000 UTC m=+2329.364117363" watchObservedRunningTime="2025-10-11 04:30:01.401123575 +0000 UTC m=+2329.364332508" Oct 11 04:30:02 crc kubenswrapper[4967]: I1011 04:30:02.395451 4967 generic.go:334] "Generic (PLEG): container finished" podID="eef2317b-924a-44b7-9fb8-0913e229df0f" containerID="0aa35a0c8da4feac2d3f08e11710e03ce2ccb39337139f9edb9878e27badd9ab" exitCode=0 Oct 11 04:30:02 crc kubenswrapper[4967]: I1011 04:30:02.395549 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-whkkp" event={"ID":"eef2317b-924a-44b7-9fb8-0913e229df0f","Type":"ContainerDied","Data":"0aa35a0c8da4feac2d3f08e11710e03ce2ccb39337139f9edb9878e27badd9ab"} Oct 11 04:30:03 crc kubenswrapper[4967]: I1011 04:30:03.841381 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-whkkp" Oct 11 04:30:03 crc kubenswrapper[4967]: I1011 04:30:03.985101 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gxd98\" (UniqueName: \"kubernetes.io/projected/eef2317b-924a-44b7-9fb8-0913e229df0f-kube-api-access-gxd98\") pod \"eef2317b-924a-44b7-9fb8-0913e229df0f\" (UID: \"eef2317b-924a-44b7-9fb8-0913e229df0f\") " Oct 11 04:30:03 crc kubenswrapper[4967]: I1011 04:30:03.985550 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eef2317b-924a-44b7-9fb8-0913e229df0f-secret-volume\") pod \"eef2317b-924a-44b7-9fb8-0913e229df0f\" (UID: \"eef2317b-924a-44b7-9fb8-0913e229df0f\") " Oct 11 04:30:03 crc kubenswrapper[4967]: I1011 04:30:03.985602 4967 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eef2317b-924a-44b7-9fb8-0913e229df0f-config-volume\") pod \"eef2317b-924a-44b7-9fb8-0913e229df0f\" (UID: \"eef2317b-924a-44b7-9fb8-0913e229df0f\") " Oct 11 04:30:03 crc kubenswrapper[4967]: I1011 04:30:03.986611 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/eef2317b-924a-44b7-9fb8-0913e229df0f-config-volume" (OuterVolumeSpecName: "config-volume") pod "eef2317b-924a-44b7-9fb8-0913e229df0f" (UID: "eef2317b-924a-44b7-9fb8-0913e229df0f"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Oct 11 04:30:03 crc kubenswrapper[4967]: I1011 04:30:03.991119 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/eef2317b-924a-44b7-9fb8-0913e229df0f-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "eef2317b-924a-44b7-9fb8-0913e229df0f" (UID: "eef2317b-924a-44b7-9fb8-0913e229df0f"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Oct 11 04:30:03 crc kubenswrapper[4967]: I1011 04:30:03.992751 4967 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/eef2317b-924a-44b7-9fb8-0913e229df0f-kube-api-access-gxd98" (OuterVolumeSpecName: "kube-api-access-gxd98") pod "eef2317b-924a-44b7-9fb8-0913e229df0f" (UID: "eef2317b-924a-44b7-9fb8-0913e229df0f"). InnerVolumeSpecName "kube-api-access-gxd98". PluginName "kubernetes.io/projected", VolumeGidValue "" Oct 11 04:30:04 crc kubenswrapper[4967]: I1011 04:30:04.087353 4967 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/eef2317b-924a-44b7-9fb8-0913e229df0f-secret-volume\") on node \"crc\" DevicePath \"\"" Oct 11 04:30:04 crc kubenswrapper[4967]: I1011 04:30:04.087392 4967 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/eef2317b-924a-44b7-9fb8-0913e229df0f-config-volume\") on node \"crc\" DevicePath \"\"" Oct 11 04:30:04 crc kubenswrapper[4967]: I1011 04:30:04.087406 4967 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gxd98\" (UniqueName: \"kubernetes.io/projected/eef2317b-924a-44b7-9fb8-0913e229df0f-kube-api-access-gxd98\") on node \"crc\" DevicePath \"\"" Oct 11 04:30:04 crc kubenswrapper[4967]: I1011 04:30:04.418682 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-whkkp" event={"ID":"eef2317b-924a-44b7-9fb8-0913e229df0f","Type":"ContainerDied","Data":"6bcd41998f5b738a32fb9b108e2462d8a01d42efada042b50a8e7aacc7c91fcf"} Oct 11 04:30:04 crc kubenswrapper[4967]: I1011 04:30:04.418744 4967 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="6bcd41998f5b738a32fb9b108e2462d8a01d42efada042b50a8e7aacc7c91fcf" Oct 11 04:30:04 crc kubenswrapper[4967]: I1011 04:30:04.418806 4967 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29335950-whkkp" Oct 11 04:30:04 crc kubenswrapper[4967]: I1011 04:30:04.496243 4967 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335905-kdpld"] Oct 11 04:30:04 crc kubenswrapper[4967]: I1011 04:30:04.504992 4967 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29335905-kdpld"] Oct 11 04:30:04 crc kubenswrapper[4967]: I1011 04:30:04.830943 4967 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c506a84b-4ea6-477e-88de-fceec3a95f31" path="/var/lib/kubelet/pods/c506a84b-4ea6-477e-88de-fceec3a95f31/volumes" Oct 11 04:30:06 crc kubenswrapper[4967]: I1011 04:30:06.814813 4967 scope.go:117] "RemoveContainer" containerID="dfd2cbbb285a882d3d3b0918a1f4268e32c2905a0b9d3bb49ba405a42cb0df28" Oct 11 04:30:06 crc kubenswrapper[4967]: E1011 04:30:06.815434 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:30:19 crc kubenswrapper[4967]: I1011 04:30:19.815245 4967 scope.go:117] "RemoveContainer" containerID="dfd2cbbb285a882d3d3b0918a1f4268e32c2905a0b9d3bb49ba405a42cb0df28" Oct 11 04:30:19 crc kubenswrapper[4967]: E1011 04:30:19.820985 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" Oct 11 04:30:22 crc kubenswrapper[4967]: I1011 04:30:22.187879 4967 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-mnt29"] Oct 11 04:30:22 crc kubenswrapper[4967]: E1011 04:30:22.188917 4967 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="eef2317b-924a-44b7-9fb8-0913e229df0f" containerName="collect-profiles" Oct 11 04:30:22 crc kubenswrapper[4967]: I1011 04:30:22.188940 4967 state_mem.go:107] "Deleted CPUSet assignment" podUID="eef2317b-924a-44b7-9fb8-0913e229df0f" containerName="collect-profiles" Oct 11 04:30:22 crc kubenswrapper[4967]: I1011 04:30:22.189279 4967 memory_manager.go:354] "RemoveStaleState removing state" podUID="eef2317b-924a-44b7-9fb8-0913e229df0f" containerName="collect-profiles" Oct 11 04:30:22 crc kubenswrapper[4967]: I1011 04:30:22.191761 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mnt29" Oct 11 04:30:22 crc kubenswrapper[4967]: I1011 04:30:22.213870 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mnt29"] Oct 11 04:30:22 crc kubenswrapper[4967]: I1011 04:30:22.315992 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f293399-2f6a-4d93-9fe6-3f5fd7a7073e-catalog-content\") pod \"certified-operators-mnt29\" (UID: \"4f293399-2f6a-4d93-9fe6-3f5fd7a7073e\") " pod="openshift-marketplace/certified-operators-mnt29" Oct 11 04:30:22 crc kubenswrapper[4967]: I1011 04:30:22.316100 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jljx2\" (UniqueName: \"kubernetes.io/projected/4f293399-2f6a-4d93-9fe6-3f5fd7a7073e-kube-api-access-jljx2\") pod \"certified-operators-mnt29\" (UID: \"4f293399-2f6a-4d93-9fe6-3f5fd7a7073e\") " pod="openshift-marketplace/certified-operators-mnt29" Oct 11 04:30:22 crc kubenswrapper[4967]: I1011 04:30:22.316437 4967 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f293399-2f6a-4d93-9fe6-3f5fd7a7073e-utilities\") pod \"certified-operators-mnt29\" (UID: \"4f293399-2f6a-4d93-9fe6-3f5fd7a7073e\") " pod="openshift-marketplace/certified-operators-mnt29" Oct 11 04:30:22 crc kubenswrapper[4967]: I1011 04:30:22.418807 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f293399-2f6a-4d93-9fe6-3f5fd7a7073e-utilities\") pod \"certified-operators-mnt29\" (UID: \"4f293399-2f6a-4d93-9fe6-3f5fd7a7073e\") " pod="openshift-marketplace/certified-operators-mnt29" Oct 11 04:30:22 crc kubenswrapper[4967]: I1011 04:30:22.418922 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f293399-2f6a-4d93-9fe6-3f5fd7a7073e-catalog-content\") pod \"certified-operators-mnt29\" (UID: \"4f293399-2f6a-4d93-9fe6-3f5fd7a7073e\") " pod="openshift-marketplace/certified-operators-mnt29" Oct 11 04:30:22 crc kubenswrapper[4967]: I1011 04:30:22.418952 4967 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jljx2\" (UniqueName: \"kubernetes.io/projected/4f293399-2f6a-4d93-9fe6-3f5fd7a7073e-kube-api-access-jljx2\") pod \"certified-operators-mnt29\" (UID: \"4f293399-2f6a-4d93-9fe6-3f5fd7a7073e\") " pod="openshift-marketplace/certified-operators-mnt29" Oct 11 04:30:22 crc kubenswrapper[4967]: I1011 04:30:22.419422 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/4f293399-2f6a-4d93-9fe6-3f5fd7a7073e-catalog-content\") pod \"certified-operators-mnt29\" (UID: \"4f293399-2f6a-4d93-9fe6-3f5fd7a7073e\") " pod="openshift-marketplace/certified-operators-mnt29" Oct 11 04:30:22 crc kubenswrapper[4967]: I1011 04:30:22.419437 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/4f293399-2f6a-4d93-9fe6-3f5fd7a7073e-utilities\") pod \"certified-operators-mnt29\" (UID: \"4f293399-2f6a-4d93-9fe6-3f5fd7a7073e\") " pod="openshift-marketplace/certified-operators-mnt29" Oct 11 04:30:22 crc kubenswrapper[4967]: I1011 04:30:22.436779 4967 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jljx2\" (UniqueName: \"kubernetes.io/projected/4f293399-2f6a-4d93-9fe6-3f5fd7a7073e-kube-api-access-jljx2\") pod \"certified-operators-mnt29\" (UID: \"4f293399-2f6a-4d93-9fe6-3f5fd7a7073e\") " pod="openshift-marketplace/certified-operators-mnt29" Oct 11 04:30:22 crc kubenswrapper[4967]: I1011 04:30:22.558738 4967 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-mnt29" Oct 11 04:30:22 crc kubenswrapper[4967]: I1011 04:30:22.827317 4967 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-mnt29"] Oct 11 04:30:23 crc kubenswrapper[4967]: I1011 04:30:23.659458 4967 generic.go:334] "Generic (PLEG): container finished" podID="4f293399-2f6a-4d93-9fe6-3f5fd7a7073e" containerID="b058ddd0645135affa726854061b07c84a4dba090fa6bbc4ebb1bf4933a4a1d6" exitCode=0 Oct 11 04:30:23 crc kubenswrapper[4967]: I1011 04:30:23.659549 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mnt29" event={"ID":"4f293399-2f6a-4d93-9fe6-3f5fd7a7073e","Type":"ContainerDied","Data":"b058ddd0645135affa726854061b07c84a4dba090fa6bbc4ebb1bf4933a4a1d6"} Oct 11 04:30:23 crc kubenswrapper[4967]: I1011 04:30:23.659915 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mnt29" event={"ID":"4f293399-2f6a-4d93-9fe6-3f5fd7a7073e","Type":"ContainerStarted","Data":"c02752e7a6de9009569399d3a35d8126829dc20f426ca8aa055264141afe73f8"} Oct 11 04:30:23 crc kubenswrapper[4967]: I1011 04:30:23.663460 4967 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Oct 11 04:30:24 crc kubenswrapper[4967]: I1011 04:30:24.675439 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mnt29" event={"ID":"4f293399-2f6a-4d93-9fe6-3f5fd7a7073e","Type":"ContainerStarted","Data":"79c685b1866ffc8dc66e6baad05427391f1998ca3696d3222dc03054ea9264d9"} Oct 11 04:30:25 crc kubenswrapper[4967]: I1011 04:30:25.689595 4967 generic.go:334] "Generic (PLEG): container finished" podID="4f293399-2f6a-4d93-9fe6-3f5fd7a7073e" containerID="79c685b1866ffc8dc66e6baad05427391f1998ca3696d3222dc03054ea9264d9" exitCode=0 Oct 11 04:30:25 crc kubenswrapper[4967]: I1011 04:30:25.689671 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mnt29" event={"ID":"4f293399-2f6a-4d93-9fe6-3f5fd7a7073e","Type":"ContainerDied","Data":"79c685b1866ffc8dc66e6baad05427391f1998ca3696d3222dc03054ea9264d9"} Oct 11 04:30:26 crc kubenswrapper[4967]: I1011 04:30:26.021767 4967 scope.go:117] "RemoveContainer" containerID="cc575196dfaaf2394c082a5d499f0ffaae1a90e9adfdc0626c04af2ad9f0fcba" Oct 11 04:30:26 crc kubenswrapper[4967]: I1011 04:30:26.068129 4967 scope.go:117] "RemoveContainer" containerID="6fcdda454c0c7f453665551fd9686dbd65bfe089b711fdd55bc4b7b27ac472f7" Oct 11 04:30:26 crc kubenswrapper[4967]: I1011 04:30:26.151244 4967 scope.go:117] "RemoveContainer" containerID="d01ccfc202d6e2c3e55352f3ed8bf385f61af024595e9dee56cb5a918d76cd6c" Oct 11 04:30:26 crc kubenswrapper[4967]: I1011 04:30:26.713206 4967 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-mnt29" event={"ID":"4f293399-2f6a-4d93-9fe6-3f5fd7a7073e","Type":"ContainerStarted","Data":"547582558ad6ef663fcfede34600bf930ca432f8822b0b1d066e2b64f8f517e4"} Oct 11 04:30:26 crc kubenswrapper[4967]: I1011 04:30:26.757105 4967 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-mnt29" podStartSLOduration=2.268095724 podStartE2EDuration="4.757062096s" podCreationTimestamp="2025-10-11 04:30:22 +0000 UTC" firstStartedPulling="2025-10-11 04:30:23.662925851 +0000 UTC m=+2351.626134824" lastFinishedPulling="2025-10-11 04:30:26.151892263 +0000 UTC m=+2354.115101196" observedRunningTime="2025-10-11 04:30:26.754230756 +0000 UTC m=+2354.717439739" watchObservedRunningTime="2025-10-11 04:30:26.757062096 +0000 UTC m=+2354.720271039" Oct 11 04:30:31 crc kubenswrapper[4967]: I1011 04:30:31.815350 4967 scope.go:117] "RemoveContainer" containerID="dfd2cbbb285a882d3d3b0918a1f4268e32c2905a0b9d3bb49ba405a42cb0df28" Oct 11 04:30:31 crc kubenswrapper[4967]: E1011 04:30:31.817114 4967 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"machine-config-daemon\" with CrashLoopBackOff: \"back-off 5m0s restarting failed container=machine-config-daemon pod=machine-config-daemon-sgrv2_openshift-machine-config-operator(97c07678-14be-410c-b61f-498cb49bc960)\"" pod="openshift-machine-config-operator/machine-config-daemon-sgrv2" podUID="97c07678-14be-410c-b61f-498cb49bc960" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515072356555024462 0ustar coreroot‹íÁ  ÷Om7 €7šÞ'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015072356556017400 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015072351570016512 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015072351570015462 5ustar corecore